var/home/core/zuul-output/0000755000175000017500000000000015115622214014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015115634141015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005247600215115634133017705 0ustar rootrootDec 08 19:21:56 crc systemd[1]: Starting Kubernetes Kubelet... Dec 08 19:21:56 crc restorecon[4691]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 19:21:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 19:21:57 crc restorecon[4691]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 08 19:21:57 crc kubenswrapper[4706]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 08 19:21:57 crc kubenswrapper[4706]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 08 19:21:57 crc kubenswrapper[4706]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 08 19:21:57 crc kubenswrapper[4706]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 08 19:21:57 crc kubenswrapper[4706]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 08 19:21:57 crc kubenswrapper[4706]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.462082 4706 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465489 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465517 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465526 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465532 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465538 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465543 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465549 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465556 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465561 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465566 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465585 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465591 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465597 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465601 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465606 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465610 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465615 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465619 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465624 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465633 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465638 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465643 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465647 4706 feature_gate.go:330] unrecognized feature gate: Example Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465652 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465656 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465661 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465666 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465670 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465674 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465679 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465683 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465688 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465692 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465697 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465702 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465708 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465715 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465720 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465726 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465732 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465737 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465742 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465746 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465751 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465755 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465760 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465772 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465777 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465783 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465789 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465794 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465799 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465804 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465808 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465813 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465820 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465825 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465830 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465834 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465839 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465844 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465848 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465852 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465856 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465860 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465865 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465871 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465877 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465882 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465888 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.465893 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466205 4706 flags.go:64] FLAG: --address="0.0.0.0" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466220 4706 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466235 4706 flags.go:64] FLAG: --anonymous-auth="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466242 4706 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466248 4706 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466254 4706 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466278 4706 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466284 4706 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466290 4706 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466295 4706 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466301 4706 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466315 4706 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466321 4706 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466327 4706 flags.go:64] FLAG: --cgroup-root="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466331 4706 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466337 4706 flags.go:64] FLAG: --client-ca-file="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466342 4706 flags.go:64] FLAG: --cloud-config="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466347 4706 flags.go:64] FLAG: --cloud-provider="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466351 4706 flags.go:64] FLAG: --cluster-dns="[]" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466359 4706 flags.go:64] FLAG: --cluster-domain="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466364 4706 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466369 4706 flags.go:64] FLAG: --config-dir="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466375 4706 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466380 4706 flags.go:64] FLAG: --container-log-max-files="5" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466387 4706 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466393 4706 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466398 4706 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466403 4706 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466408 4706 flags.go:64] FLAG: --contention-profiling="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466414 4706 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466420 4706 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466425 4706 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466432 4706 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466439 4706 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466444 4706 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466450 4706 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466455 4706 flags.go:64] FLAG: --enable-load-reader="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466460 4706 flags.go:64] FLAG: --enable-server="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466465 4706 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466471 4706 flags.go:64] FLAG: --event-burst="100" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466476 4706 flags.go:64] FLAG: --event-qps="50" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466482 4706 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466487 4706 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466492 4706 flags.go:64] FLAG: --eviction-hard="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466499 4706 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466504 4706 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466509 4706 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466520 4706 flags.go:64] FLAG: --eviction-soft="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466526 4706 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466531 4706 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466536 4706 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466541 4706 flags.go:64] FLAG: --experimental-mounter-path="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466546 4706 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466551 4706 flags.go:64] FLAG: --fail-swap-on="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466556 4706 flags.go:64] FLAG: --feature-gates="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466569 4706 flags.go:64] FLAG: --file-check-frequency="20s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466575 4706 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466580 4706 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466585 4706 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466592 4706 flags.go:64] FLAG: --healthz-port="10248" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466598 4706 flags.go:64] FLAG: --help="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466604 4706 flags.go:64] FLAG: --hostname-override="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466610 4706 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466616 4706 flags.go:64] FLAG: --http-check-frequency="20s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466622 4706 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466627 4706 flags.go:64] FLAG: --image-credential-provider-config="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466632 4706 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466637 4706 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466642 4706 flags.go:64] FLAG: --image-service-endpoint="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466647 4706 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466652 4706 flags.go:64] FLAG: --kube-api-burst="100" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466657 4706 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466663 4706 flags.go:64] FLAG: --kube-api-qps="50" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466668 4706 flags.go:64] FLAG: --kube-reserved="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466673 4706 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466678 4706 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466683 4706 flags.go:64] FLAG: --kubelet-cgroups="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466688 4706 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466693 4706 flags.go:64] FLAG: --lock-file="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466698 4706 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466703 4706 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466708 4706 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466717 4706 flags.go:64] FLAG: --log-json-split-stream="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466731 4706 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466736 4706 flags.go:64] FLAG: --log-text-split-stream="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466741 4706 flags.go:64] FLAG: --logging-format="text" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466746 4706 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466752 4706 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466757 4706 flags.go:64] FLAG: --manifest-url="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466762 4706 flags.go:64] FLAG: --manifest-url-header="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466769 4706 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466774 4706 flags.go:64] FLAG: --max-open-files="1000000" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466781 4706 flags.go:64] FLAG: --max-pods="110" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466788 4706 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466794 4706 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466799 4706 flags.go:64] FLAG: --memory-manager-policy="None" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466805 4706 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466810 4706 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466815 4706 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466820 4706 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466834 4706 flags.go:64] FLAG: --node-status-max-images="50" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466839 4706 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466844 4706 flags.go:64] FLAG: --oom-score-adj="-999" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466849 4706 flags.go:64] FLAG: --pod-cidr="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466854 4706 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466864 4706 flags.go:64] FLAG: --pod-manifest-path="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466869 4706 flags.go:64] FLAG: --pod-max-pids="-1" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466875 4706 flags.go:64] FLAG: --pods-per-core="0" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466881 4706 flags.go:64] FLAG: --port="10250" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466886 4706 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466892 4706 flags.go:64] FLAG: --provider-id="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466897 4706 flags.go:64] FLAG: --qos-reserved="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466902 4706 flags.go:64] FLAG: --read-only-port="10255" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466907 4706 flags.go:64] FLAG: --register-node="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466912 4706 flags.go:64] FLAG: --register-schedulable="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466917 4706 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466927 4706 flags.go:64] FLAG: --registry-burst="10" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466932 4706 flags.go:64] FLAG: --registry-qps="5" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466937 4706 flags.go:64] FLAG: --reserved-cpus="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466946 4706 flags.go:64] FLAG: --reserved-memory="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466953 4706 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466959 4706 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466964 4706 flags.go:64] FLAG: --rotate-certificates="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466970 4706 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466976 4706 flags.go:64] FLAG: --runonce="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466981 4706 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466988 4706 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466994 4706 flags.go:64] FLAG: --seccomp-default="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.466999 4706 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467005 4706 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467011 4706 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467016 4706 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467021 4706 flags.go:64] FLAG: --storage-driver-password="root" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467027 4706 flags.go:64] FLAG: --storage-driver-secure="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467033 4706 flags.go:64] FLAG: --storage-driver-table="stats" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467038 4706 flags.go:64] FLAG: --storage-driver-user="root" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467043 4706 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467048 4706 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467053 4706 flags.go:64] FLAG: --system-cgroups="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467059 4706 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467067 4706 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467073 4706 flags.go:64] FLAG: --tls-cert-file="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467078 4706 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467086 4706 flags.go:64] FLAG: --tls-min-version="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467091 4706 flags.go:64] FLAG: --tls-private-key-file="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467097 4706 flags.go:64] FLAG: --topology-manager-policy="none" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467102 4706 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467107 4706 flags.go:64] FLAG: --topology-manager-scope="container" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467112 4706 flags.go:64] FLAG: --v="2" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467119 4706 flags.go:64] FLAG: --version="false" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467126 4706 flags.go:64] FLAG: --vmodule="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467133 4706 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467139 4706 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467316 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467326 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467333 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467338 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467343 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467349 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467354 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467359 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467364 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467368 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467374 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467379 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467384 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467389 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467394 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467399 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467404 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467408 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467413 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467417 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467422 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467426 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467431 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467435 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467440 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467446 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467451 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467456 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467460 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467465 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467469 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467473 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467479 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467485 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467490 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467494 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467500 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467506 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467511 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467516 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467521 4706 feature_gate.go:330] unrecognized feature gate: Example Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467532 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467538 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467543 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467548 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467553 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467557 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467561 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467565 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467569 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467572 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467576 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467579 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467582 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467586 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467590 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467594 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467599 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467603 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467607 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467611 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467615 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467619 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467622 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467626 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467629 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467633 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467636 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467640 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467646 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.467650 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.467657 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.476836 4706 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.476856 4706 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476930 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476936 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476940 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476945 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476949 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476954 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476961 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476965 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476969 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476974 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476979 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476983 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476988 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476991 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476995 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.476999 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477003 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477007 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477011 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477015 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477019 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477023 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477026 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477030 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477033 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477037 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477040 4706 feature_gate.go:330] unrecognized feature gate: Example Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477044 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477049 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477054 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477058 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477062 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477065 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477069 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477072 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477077 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477082 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477086 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477089 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477093 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477096 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477100 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477104 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477108 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477111 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477115 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477118 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477122 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477125 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477129 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477132 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477136 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477139 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477143 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477146 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477150 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477153 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477157 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477160 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477164 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477168 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477172 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477175 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477179 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477182 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477187 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477191 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477194 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477198 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477201 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477205 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.477211 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477340 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477346 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477350 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477354 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477358 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477362 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477367 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477371 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477374 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477378 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477381 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477385 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477388 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477392 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477395 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477399 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477404 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477410 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477415 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477419 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477423 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477427 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477431 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477435 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477440 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477446 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477450 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477454 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477458 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477462 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477466 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477471 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477475 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477480 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477485 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477490 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477494 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477498 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477502 4706 feature_gate.go:330] unrecognized feature gate: Example Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477506 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477510 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477514 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477518 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477522 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477525 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477528 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477532 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477536 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477539 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477542 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477547 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477553 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477557 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477562 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477566 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477569 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477574 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477578 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477581 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477585 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477589 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477593 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477596 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477600 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477604 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477608 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477611 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477615 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477619 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477623 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.477626 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.477632 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.478314 4706 server.go:940] "Client rotation is on, will bootstrap in background" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.483328 4706 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.483439 4706 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.484006 4706 server.go:997] "Starting client certificate rotation" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.484032 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.484595 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-08 22:27:12.486160649 +0000 UTC Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.484821 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.488892 4706 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.490322 4706 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.491237 4706 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.499798 4706 log.go:25] "Validated CRI v1 runtime API" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.512425 4706 log.go:25] "Validated CRI v1 image API" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.514237 4706 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.516923 4706 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-08-19-17-38-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.516969 4706 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.536751 4706 manager.go:217] Machine: {Timestamp:2025-12-08 19:21:57.533813616 +0000 UTC m=+0.176014639 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:fcc1c5cb-e33f-45eb-b8f8-156e819c7d80 BootID:cf23845b-400b-4cbc-a718-15bd67773250 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f3:5e:85 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f3:5e:85 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:79:58:dd Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:27:91:be Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:98:43:df Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:7e:f0:85 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:1e:94:9e:bd:e4:22 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:0a:c9:a4:14:3b:32 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.537059 4706 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.537386 4706 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.537825 4706 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.538115 4706 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.538165 4706 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.538465 4706 topology_manager.go:138] "Creating topology manager with none policy" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.538483 4706 container_manager_linux.go:303] "Creating device plugin manager" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.538722 4706 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.538779 4706 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.539033 4706 state_mem.go:36] "Initialized new in-memory state store" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.539159 4706 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.539808 4706 kubelet.go:418] "Attempting to sync node with API server" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.539832 4706 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.539877 4706 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.539898 4706 kubelet.go:324] "Adding apiserver pod source" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.539918 4706 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.542078 4706 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.542292 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.542426 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.542433 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.542505 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.542560 4706 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.543432 4706 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544063 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544093 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544102 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544111 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544134 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544144 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544153 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544167 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544176 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544185 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544220 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544234 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.544472 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.545014 4706 server.go:1280] "Started kubelet" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.545641 4706 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.545827 4706 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 08 19:21:57 crc systemd[1]: Started Kubernetes Kubelet. Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.547962 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.548025 4706 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.547795 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.548416 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 08:15:09.052675375 +0000 UTC Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.548518 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 492h53m11.504160323s for next certificate rotation Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.548545 4706 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.548649 4706 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.548560 4706 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.549151 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.549229 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.549791 4706 factory.go:55] Registering systemd factory Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.549821 4706 factory.go:221] Registration of the systemd container factory successfully Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.548998 4706 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.550354 4706 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.550891 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="200ms" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.550332 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f53ccb012e73a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-08 19:21:57.544986426 +0000 UTC m=+0.187187439,LastTimestamp:2025-12-08 19:21:57.544986426 +0000 UTC m=+0.187187439,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.551563 4706 factory.go:153] Registering CRI-O factory Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.551627 4706 factory.go:221] Registration of the crio container factory successfully Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.551808 4706 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.551861 4706 factory.go:103] Registering Raw factory Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.551987 4706 manager.go:1196] Started watching for new ooms in manager Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.555643 4706 manager.go:319] Starting recovery of all containers Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.560962 4706 server.go:460] "Adding debug handlers to kubelet server" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563572 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563673 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563691 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563707 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563722 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563771 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563788 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563801 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563821 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563838 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563855 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563870 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563885 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563922 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563935 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563950 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563966 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.563999 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564012 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564025 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564038 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564051 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564067 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564085 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564105 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564128 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564149 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564170 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564189 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564218 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564235 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564251 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564294 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564310 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564353 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564369 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564384 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564399 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564413 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564429 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564449 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564491 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564506 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564523 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564540 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564575 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564594 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564613 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564633 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564652 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564667 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564684 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564708 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564725 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564749 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564768 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564786 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564802 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564818 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564833 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564848 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564867 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564883 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564898 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564915 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564931 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564947 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564961 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564977 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.564993 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565006 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565021 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565036 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565051 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565066 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565083 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565101 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565123 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565143 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565159 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565174 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565190 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565216 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565235 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565256 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565303 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565322 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565337 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565355 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565371 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565410 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565424 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565438 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.565456 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566321 4706 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566363 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566391 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566418 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566444 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566466 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566486 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566516 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566535 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566555 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566575 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566604 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566629 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566657 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566674 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566690 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566706 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566724 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566739 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566756 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566775 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566792 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566807 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566826 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566868 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566884 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566902 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566919 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566933 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566968 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.566982 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567003 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567020 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567034 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567048 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567064 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567080 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567096 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567111 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567125 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567139 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567155 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567170 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567188 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567204 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567221 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567236 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567254 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567288 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567304 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567317 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567333 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567350 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567368 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567383 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567421 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567434 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567448 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567462 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567476 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567491 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567506 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567522 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567538 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567560 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567581 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567603 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567625 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567639 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567653 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567666 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567682 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567698 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567761 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567775 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567788 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567801 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567816 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567850 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567865 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567879 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567895 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567908 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567924 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567937 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567953 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.567988 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568003 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568017 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568033 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568047 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568064 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568082 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568100 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568115 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568130 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568146 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568166 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568186 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568210 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568232 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568247 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568282 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568297 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568312 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568331 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568347 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568359 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568372 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568386 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568400 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568416 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568430 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568451 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568464 4706 reconstruct.go:97] "Volume reconstruction finished" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.568474 4706 reconciler.go:26] "Reconciler: start to sync state" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.580438 4706 manager.go:324] Recovery completed Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.591569 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.593444 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.593498 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.593509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.595289 4706 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.595306 4706 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.595331 4706 state_mem.go:36] "Initialized new in-memory state store" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.604463 4706 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.605614 4706 policy_none.go:49] "None policy: Start" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.606553 4706 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.606588 4706 state_mem.go:35] "Initializing new in-memory state store" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.606608 4706 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.607010 4706 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.607037 4706 kubelet.go:2335] "Starting kubelet main sync loop" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.607077 4706 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 08 19:21:57 crc kubenswrapper[4706]: W1208 19:21:57.608013 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.608093 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.650884 4706 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.657531 4706 manager.go:334] "Starting Device Plugin manager" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.657817 4706 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.657834 4706 server.go:79] "Starting device plugin registration server" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.658504 4706 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.658530 4706 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.658739 4706 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.658865 4706 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.658879 4706 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.665538 4706 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.707474 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.707648 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.709138 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.709191 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.709203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.709414 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.709790 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.709824 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710368 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710414 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710428 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710508 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710621 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710758 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.710802 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.711591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.711619 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.711632 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.711637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.711656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.711667 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.711744 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.711944 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.712001 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.712652 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.712683 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.712698 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.712821 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.712967 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713018 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.712979 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713676 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713703 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713828 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713897 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.713926 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.714490 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.714509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.714519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.751505 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="400ms" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.758872 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.760041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.760100 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.760114 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.760147 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.760865 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771026 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771055 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771077 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771094 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771120 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771178 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771222 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771240 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771288 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771382 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771439 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771465 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771494 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771522 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.771536 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.872891 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.872952 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.872971 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.872989 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873006 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873022 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873038 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873054 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873072 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873088 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873108 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873125 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873141 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873227 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873244 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873292 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873238 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873159 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873146 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873376 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873199 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873172 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873496 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873671 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873714 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873746 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873868 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.873816 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.961695 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.962989 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.963079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.963094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:57 crc kubenswrapper[4706]: I1208 19:21:57.963123 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 19:21:57 crc kubenswrapper[4706]: E1208 19:21:57.963628 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.033826 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.040757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.057931 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:21:58 crc kubenswrapper[4706]: W1208 19:21:58.059972 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-148757d14334745b3dc3833348064c294f07cdd9e993af6bba32e85b1fe0cc66 WatchSource:0}: Error finding container 148757d14334745b3dc3833348064c294f07cdd9e993af6bba32e85b1fe0cc66: Status 404 returned error can't find the container with id 148757d14334745b3dc3833348064c294f07cdd9e993af6bba32e85b1fe0cc66 Dec 08 19:21:58 crc kubenswrapper[4706]: W1208 19:21:58.065428 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-da7b7d49d5a60d01f90481af645b3efbb14b64759641f525c74bf31dcbdc9670 WatchSource:0}: Error finding container da7b7d49d5a60d01f90481af645b3efbb14b64759641f525c74bf31dcbdc9670: Status 404 returned error can't find the container with id da7b7d49d5a60d01f90481af645b3efbb14b64759641f525c74bf31dcbdc9670 Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.080992 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:21:58 crc kubenswrapper[4706]: W1208 19:21:58.081342 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-80ee47027783cffd30c5724578ba296e9be3bae2ae019d084dd97de2e14ff116 WatchSource:0}: Error finding container 80ee47027783cffd30c5724578ba296e9be3bae2ae019d084dd97de2e14ff116: Status 404 returned error can't find the container with id 80ee47027783cffd30c5724578ba296e9be3bae2ae019d084dd97de2e14ff116 Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.086655 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:21:58 crc kubenswrapper[4706]: W1208 19:21:58.102157 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-09fc47ac5641fe86afbd9d5caedf3650b6d7544650510d0c980eb5c5ab65136d WatchSource:0}: Error finding container 09fc47ac5641fe86afbd9d5caedf3650b6d7544650510d0c980eb5c5ab65136d: Status 404 returned error can't find the container with id 09fc47ac5641fe86afbd9d5caedf3650b6d7544650510d0c980eb5c5ab65136d Dec 08 19:21:58 crc kubenswrapper[4706]: W1208 19:21:58.107601 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-576cedfc894dd7ad803048a30fea29d0dc542f2bc85737c778c300c10c7479cb WatchSource:0}: Error finding container 576cedfc894dd7ad803048a30fea29d0dc542f2bc85737c778c300c10c7479cb: Status 404 returned error can't find the container with id 576cedfc894dd7ad803048a30fea29d0dc542f2bc85737c778c300c10c7479cb Dec 08 19:21:58 crc kubenswrapper[4706]: E1208 19:21:58.152958 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="800ms" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.364193 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.366862 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.366933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.366953 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.366992 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 19:21:58 crc kubenswrapper[4706]: E1208 19:21:58.367658 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Dec 08 19:21:58 crc kubenswrapper[4706]: W1208 19:21:58.389771 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:58 crc kubenswrapper[4706]: E1208 19:21:58.389842 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:58 crc kubenswrapper[4706]: W1208 19:21:58.469953 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:58 crc kubenswrapper[4706]: E1208 19:21:58.470077 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.549617 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.612044 4706 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338" exitCode=0 Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.612150 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.612297 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"576cedfc894dd7ad803048a30fea29d0dc542f2bc85737c778c300c10c7479cb"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.612437 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.614466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.614517 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.614530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.615444 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.615532 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"09fc47ac5641fe86afbd9d5caedf3650b6d7544650510d0c980eb5c5ab65136d"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.617767 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb" exitCode=0 Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.617819 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.617866 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"80ee47027783cffd30c5724578ba296e9be3bae2ae019d084dd97de2e14ff116"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.617961 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.619031 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.619071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.619085 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.619865 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f" exitCode=0 Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.619904 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.619943 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"da7b7d49d5a60d01f90481af645b3efbb14b64759641f525c74bf31dcbdc9670"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.620032 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.620693 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621205 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621817 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644" exitCode=0 Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621837 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621851 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621871 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"148757d14334745b3dc3833348064c294f07cdd9e993af6bba32e85b1fe0cc66"} Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.621974 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.622626 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.622657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:58 crc kubenswrapper[4706]: I1208 19:21:58.622668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:58 crc kubenswrapper[4706]: W1208 19:21:58.849314 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:58 crc kubenswrapper[4706]: E1208 19:21:58.849879 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:58 crc kubenswrapper[4706]: E1208 19:21:58.953709 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="1.6s" Dec 08 19:21:59 crc kubenswrapper[4706]: W1208 19:21:59.120531 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Dec 08 19:21:59 crc kubenswrapper[4706]: E1208 19:21:59.120659 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.169081 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.171486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.171547 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.171586 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.171648 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 19:21:59 crc kubenswrapper[4706]: E1208 19:21:59.172339 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.595064 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.627582 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.627639 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.627653 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.627789 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.629277 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.629317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.629331 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.631881 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.631917 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.631934 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.632091 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.633766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.633816 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.633832 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.634114 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.634141 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.634154 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.634164 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.638077 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1fbd3072ee050c9a83353949c6a2f3d761019bd054ae54a97ea97a9aea27fd67"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.638284 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.645702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.645756 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.645771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.647664 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab" exitCode=0 Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.647721 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab"} Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.647863 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.656990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.657106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.657122 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:21:59 crc kubenswrapper[4706]: I1208 19:21:59.793607 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.663681 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0"} Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.663781 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.665145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.665257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.665358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.668060 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb" exitCode=0 Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.668176 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb"} Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.668328 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.668404 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.668486 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670246 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670285 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670303 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670385 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.670472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.773047 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.774654 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.774700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.774710 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.774742 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 19:22:00 crc kubenswrapper[4706]: I1208 19:22:00.944924 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.450806 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.675817 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b"} Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.675912 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d"} Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.675932 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.675940 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353"} Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.675970 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2"} Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.677017 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.677095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.677121 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.814922 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.815246 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.817224 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.817360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.817391 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:01 crc kubenswrapper[4706]: I1208 19:22:01.823645 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.130555 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.688483 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1"} Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.688659 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.688789 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.688879 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.690837 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.690887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.690907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.691036 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.691036 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.691078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.691134 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.691250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:02 crc kubenswrapper[4706]: I1208 19:22:02.691313 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.209868 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.692910 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.693028 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.693130 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.693243 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695251 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695316 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695374 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695251 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695451 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:03 crc kubenswrapper[4706]: I1208 19:22:03.695416 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.423465 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.424321 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.424451 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.426121 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.426167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.426182 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.762501 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.762704 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.763981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.764055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.764069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.986775 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.987060 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.988928 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.988978 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:05 crc kubenswrapper[4706]: I1208 19:22:05.988997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:07 crc kubenswrapper[4706]: E1208 19:22:07.790395 4706 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 08 19:22:07 crc kubenswrapper[4706]: I1208 19:22:07.790932 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 08 19:22:07 crc kubenswrapper[4706]: I1208 19:22:07.791059 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:07 crc kubenswrapper[4706]: I1208 19:22:07.791951 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:07 crc kubenswrapper[4706]: I1208 19:22:07.791977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:07 crc kubenswrapper[4706]: I1208 19:22:07.791986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:08 crc kubenswrapper[4706]: I1208 19:22:08.424069 4706 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 08 19:22:08 crc kubenswrapper[4706]: I1208 19:22:08.424170 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 08 19:22:09 crc kubenswrapper[4706]: I1208 19:22:09.550592 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 08 19:22:09 crc kubenswrapper[4706]: E1208 19:22:09.597799 4706 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 08 19:22:10 crc kubenswrapper[4706]: E1208 19:22:10.555340 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 08 19:22:10 crc kubenswrapper[4706]: I1208 19:22:10.567690 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 08 19:22:10 crc kubenswrapper[4706]: I1208 19:22:10.567754 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 08 19:22:10 crc kubenswrapper[4706]: I1208 19:22:10.576466 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 08 19:22:10 crc kubenswrapper[4706]: I1208 19:22:10.576527 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.219708 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.220027 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.221839 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.221899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.221909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.225902 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.808010 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.808080 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.809453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.809646 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.809781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.851916 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 08 19:22:13 crc kubenswrapper[4706]: I1208 19:22:13.868061 4706 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.578940 4706 trace.go:236] Trace[382382871]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Dec-2025 19:22:01.323) (total time: 14255ms): Dec 08 19:22:15 crc kubenswrapper[4706]: Trace[382382871]: ---"Objects listed" error: 14255ms (19:22:15.578) Dec 08 19:22:15 crc kubenswrapper[4706]: Trace[382382871]: [14.25506757s] [14.25506757s] END Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.579644 4706 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.579574 4706 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.579096 4706 trace.go:236] Trace[2081739136]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Dec-2025 19:22:01.333) (total time: 14245ms): Dec 08 19:22:15 crc kubenswrapper[4706]: Trace[2081739136]: ---"Objects listed" error: 14245ms (19:22:15.578) Dec 08 19:22:15 crc kubenswrapper[4706]: Trace[2081739136]: [14.245589207s] [14.245589207s] END Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.579740 4706 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.579411 4706 trace.go:236] Trace[1241447506]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Dec-2025 19:22:01.157) (total time: 14421ms): Dec 08 19:22:15 crc kubenswrapper[4706]: Trace[1241447506]: ---"Objects listed" error: 14421ms (19:22:15.579) Dec 08 19:22:15 crc kubenswrapper[4706]: Trace[1241447506]: [14.421616729s] [14.421616729s] END Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.579850 4706 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 08 19:22:15 crc kubenswrapper[4706]: E1208 19:22:15.580717 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.581325 4706 trace.go:236] Trace[1559574819]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Dec-2025 19:22:00.997) (total time: 14583ms): Dec 08 19:22:15 crc kubenswrapper[4706]: Trace[1559574819]: ---"Objects listed" error: 14583ms (19:22:15.581) Dec 08 19:22:15 crc kubenswrapper[4706]: Trace[1559574819]: [14.583674267s] [14.583674267s] END Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.581350 4706 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.617456 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.622008 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.624218 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.684404 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58104->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.684469 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58104->192.168.126.11:17697: read: connection reset by peer" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.685714 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.685776 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.814980 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.816563 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0" exitCode=255 Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.816625 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0"} Dec 08 19:22:15 crc kubenswrapper[4706]: E1208 19:22:15.822602 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:22:15 crc kubenswrapper[4706]: I1208 19:22:15.823976 4706 scope.go:117] "RemoveContainer" containerID="269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.011502 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.022753 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.123628 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.550020 4706 apiserver.go:52] "Watching apiserver" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.554480 4706 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.555082 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-etcd/etcd-crc","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/iptables-alerter-4ln5h"] Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.555671 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.555769 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.555771 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.555904 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.555934 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.556069 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.556567 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.556601 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.556703 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.558600 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.558665 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.558599 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.559464 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.559705 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.560454 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.561823 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.562010 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.571032 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.590673 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.607768 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.618376 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.642060 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.649854 4706 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.659907 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.677880 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.686150 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.686227 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.686281 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.686312 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.686340 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.686364 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.686387 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.686412 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.697179 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.697367 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.697375 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.697830 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.698040 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.698258 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.698287 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.698678 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.696447 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.698827 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.698891 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.698944 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.699561 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.699931 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.699993 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700038 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700079 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700124 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700167 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700204 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700244 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700731 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700780 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700809 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.700832 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.701565 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.701578 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.701693 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.701750 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.701758 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.701804 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.701928 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702010 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702070 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702104 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702141 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702180 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702212 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702240 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702302 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702338 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702369 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702397 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702433 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702463 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702493 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702524 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702561 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702594 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702624 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702650 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702680 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702707 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702731 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702771 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702803 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702838 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702869 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702898 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702929 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.702957 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.703038 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704510 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704574 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704602 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704626 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704648 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704670 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704689 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704708 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704726 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704748 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704767 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704785 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704807 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704829 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704866 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704898 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704930 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704958 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.704985 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705039 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705067 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705099 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705129 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705152 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705175 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705197 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705218 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705247 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705296 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705320 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705344 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705364 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705385 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705412 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705435 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705457 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705481 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705506 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705533 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705555 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705576 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705599 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705617 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705638 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705728 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705755 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705780 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705804 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.701851 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705813 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.705920 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.706039 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.706223 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.706355 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.706427 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.706504 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.706553 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.706701 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.706903 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.707025 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.707057 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.707221 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708274 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708310 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708337 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708367 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708394 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708423 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708453 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709363 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709407 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709513 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709579 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709609 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709662 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709694 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709725 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709755 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709779 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709889 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710119 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710158 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710954 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710984 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711016 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711051 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711144 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711178 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711211 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711241 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711293 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711322 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711352 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711378 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711403 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711430 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711459 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711485 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711510 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711538 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711570 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711600 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711633 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711870 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712232 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712813 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712850 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712878 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712937 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712971 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713020 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713051 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713089 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713129 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713159 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713190 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713220 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713278 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713310 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713337 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713364 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713394 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713426 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713454 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713484 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713512 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713539 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713569 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713599 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713630 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713657 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713683 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713710 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713779 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713812 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713841 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713869 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713898 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714250 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714301 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714335 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714368 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714395 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714425 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714460 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714601 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714633 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714664 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714734 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714779 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714812 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714849 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714883 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715104 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715598 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715635 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715692 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715724 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715757 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715814 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715848 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715884 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715987 4706 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716006 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716022 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716036 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716051 4706 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716066 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716082 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716098 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716115 4706 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716133 4706 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716151 4706 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716166 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716182 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716196 4706 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.707320 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.707459 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.707512 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708063 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708127 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708147 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708849 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.708862 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709082 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709139 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709165 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709353 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709574 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709607 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709798 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.709874 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710039 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710159 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710292 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710691 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.710994 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711137 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711149 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711324 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711471 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711545 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711565 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711572 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711593 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.711880 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712123 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712208 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712303 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712321 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712487 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712917 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712929 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713055 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.712923 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713220 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.723886 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713460 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.713871 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714145 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714340 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.714408 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715145 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715435 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715480 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715709 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.715798 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716026 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716129 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716141 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716360 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716610 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716790 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716876 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716980 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.716930 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.717192 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.717429 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.717635 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.717720 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.717762 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.718140 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.718241 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.718252 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.718628 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.718647 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.718825 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.718840 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.719025 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.719064 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.718574 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.719123 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.719226 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.719384 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.719569 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.719872 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.720435 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.721043 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.722085 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.722200 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.722409 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.722653 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.722706 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.723211 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.723278 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.723304 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.724376 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.724703 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.724778 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.725130 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.725171 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.725810 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.725917 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.725962 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.726255 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.726256 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.726887 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.727107 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.727240 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.727409 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.727481 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.727499 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.727890 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.727909 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.727967 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728136 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728184 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728337 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728506 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728656 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728679 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728784 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728857 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.728968 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.729392 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.729641 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.730063 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.730285 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:22:17.23023661 +0000 UTC m=+19.872437623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.730202 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.730430 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.730522 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.730507 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.730650 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.730612 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.730939 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.731032 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.731052 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.731150 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.731213 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.731489 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.731696 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.731777 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.731872 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:17.231847234 +0000 UTC m=+19.874048457 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.731954 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.732140 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.732277 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.732288 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.732299 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:17.232237195 +0000 UTC m=+19.874438428 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.731959 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.732383 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.732523 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.732571 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.733093 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.733782 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.734067 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.734942 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.734984 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.735416 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.735502 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.736233 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.736635 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.737028 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.737105 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.737141 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.737165 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.737301 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:17.237250522 +0000 UTC m=+19.879451715 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.737309 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.737326 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.737422 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.739515 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.740501 4706 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.745168 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.747205 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.747772 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.747792 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.747807 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.747869 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:17.247846712 +0000 UTC m=+19.890047715 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.747919 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.748242 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.748351 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.750931 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.751720 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.751847 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.751888 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.752333 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.753100 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.753232 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.753280 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.753436 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.755512 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.757219 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.759751 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.760001 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.760085 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.760530 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.761374 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.763180 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.765731 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.771818 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.774519 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.775143 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.776071 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.778281 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.787125 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.789523 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.816736 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817061 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817162 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817221 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817372 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817483 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817555 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817642 4706 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817654 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817666 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817678 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817688 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817697 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817706 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817715 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817598 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817725 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817768 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817783 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817795 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817808 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817821 4706 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817834 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817847 4706 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817859 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817871 4706 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817882 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817897 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817910 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817919 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817928 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817938 4706 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817947 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817956 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817964 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817973 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817981 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817990 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.817999 4706 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818008 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818017 4706 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818025 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818034 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818042 4706 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818050 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818058 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818067 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818076 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818084 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818094 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818103 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818115 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818125 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818136 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818145 4706 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818154 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818163 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818172 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818180 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818188 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818198 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818207 4706 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818215 4706 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818224 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818233 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818242 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818251 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818274 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818284 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818292 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818301 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818309 4706 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818319 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818328 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818336 4706 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818344 4706 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818354 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818363 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818372 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818382 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818390 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818398 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818406 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818415 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818424 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818433 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818442 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818450 4706 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818461 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818470 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818478 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818487 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818495 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818503 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818511 4706 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818520 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818529 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818537 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818548 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818559 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818571 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818584 4706 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818596 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818606 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818618 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818630 4706 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818642 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818653 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818667 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818680 4706 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818690 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818700 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818708 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818718 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818726 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818736 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818752 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818761 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818769 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818777 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818785 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818794 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818802 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818810 4706 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818819 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818827 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818836 4706 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818844 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818852 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818861 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818869 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818877 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818886 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818894 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818903 4706 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818912 4706 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818920 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818928 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818936 4706 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818945 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818952 4706 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818961 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818970 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818977 4706 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818986 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.818995 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819003 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819011 4706 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819019 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819028 4706 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819037 4706 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819045 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819053 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819062 4706 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819073 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819083 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819093 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819103 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819113 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819124 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819134 4706 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819145 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819155 4706 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819166 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819177 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819191 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819203 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819214 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819227 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819238 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819250 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819281 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819291 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819301 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819311 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819321 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819332 4706 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819343 4706 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819354 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819366 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819378 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819389 4706 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819400 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.819409 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.823040 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.824631 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1"} Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.841061 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 08 19:22:16 crc kubenswrapper[4706]: E1208 19:22:16.841552 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.848558 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.869436 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.869836 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.879606 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.885793 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 19:22:16 crc kubenswrapper[4706]: W1208 19:22:16.902654 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-5a855e370dc64d9e7e56dbbeb91f4cfd49c89d8d16c27b2b5e8b164ba937af11 WatchSource:0}: Error finding container 5a855e370dc64d9e7e56dbbeb91f4cfd49c89d8d16c27b2b5e8b164ba937af11: Status 404 returned error can't find the container with id 5a855e370dc64d9e7e56dbbeb91f4cfd49c89d8d16c27b2b5e8b164ba937af11 Dec 08 19:22:16 crc kubenswrapper[4706]: W1208 19:22:16.903973 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-de8b15ada0561d5971a639de2b8b3c76cbed722dfc18da3dc69d35a1318c654f WatchSource:0}: Error finding container de8b15ada0561d5971a639de2b8b3c76cbed722dfc18da3dc69d35a1318c654f: Status 404 returned error can't find the container with id de8b15ada0561d5971a639de2b8b3c76cbed722dfc18da3dc69d35a1318c654f Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.912912 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.924654 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.937306 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.953900 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.972394 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:16 crc kubenswrapper[4706]: I1208 19:22:16.996933 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.008483 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.324831 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.324972 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325074 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:22:18.325036388 +0000 UTC m=+20.967237391 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325162 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.325179 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.325221 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.325251 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325191 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325332 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325432 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:18.325406388 +0000 UTC m=+20.967607431 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325452 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325469 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325483 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325564 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:18.325554552 +0000 UTC m=+20.967755555 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325639 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325906 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:18.32582475 +0000 UTC m=+20.968025893 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.325909 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.326088 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:18.326071597 +0000 UTC m=+20.968272640 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.608133 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:17 crc kubenswrapper[4706]: E1208 19:22:17.608427 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.616015 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.616790 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.618609 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.620293 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.622113 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.623347 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.624958 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.627037 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.628802 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.630566 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.630913 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.631874 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.634235 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.634962 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.635782 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.637381 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.638130 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.640577 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.641702 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.643500 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.646299 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.647549 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.649876 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.650903 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.653244 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.654164 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.655440 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.655692 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.657730 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.658880 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.660866 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.661818 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.663716 4706 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.663923 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.667406 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.668478 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.670148 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.673449 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.674917 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.676843 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.677276 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.678171 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.680654 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.681654 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.683807 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.685359 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.687570 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.688179 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.689434 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.690242 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.691662 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.692214 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.692771 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.693855 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.694179 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.694729 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.695841 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.696409 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.711982 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.726981 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.747524 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.780831 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.802627 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.844624 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6"} Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.844823 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5"} Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.844860 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5a855e370dc64d9e7e56dbbeb91f4cfd49c89d8d16c27b2b5e8b164ba937af11"} Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.848623 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c"} Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.848830 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a8d0cea3f19287e89a1a155e3d6dd93351b4457296c313034517951c3e88855c"} Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.851313 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"de8b15ada0561d5971a639de2b8b3c76cbed722dfc18da3dc69d35a1318c654f"} Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.851618 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.866388 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.892067 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.909200 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.925849 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.951329 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.970990 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:17 crc kubenswrapper[4706]: I1208 19:22:17.986217 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.000303 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.017903 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.031237 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.050860 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.068864 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.083915 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.097501 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.111585 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.152096 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.180884 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.197087 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.333956 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.334097 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.334127 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.334157 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334177 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:22:20.334153062 +0000 UTC m=+22.976354065 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.334208 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334326 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334361 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334381 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334396 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334328 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334383 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:20.334371378 +0000 UTC m=+22.976572381 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334456 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:20.33444805 +0000 UTC m=+22.976649053 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334469 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:20.33446257 +0000 UTC m=+22.976663573 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334504 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334573 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334588 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.334712 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:20.334672986 +0000 UTC m=+22.976873989 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.608080 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.608112 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.608235 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.608426 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.781768 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.784603 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.784664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.784675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.784744 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.793005 4706 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.793375 4706 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.794797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.794846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.794869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.794942 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.794954 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:18Z","lastTransitionTime":"2025-12-08T19:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.811653 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.815601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.815637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.815648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.815663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.815678 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:18Z","lastTransitionTime":"2025-12-08T19:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.828071 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.834212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.834326 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.834346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.834377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.834398 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:18Z","lastTransitionTime":"2025-12-08T19:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.851558 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.857252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.857310 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.857325 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.857344 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.857362 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:18Z","lastTransitionTime":"2025-12-08T19:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.873608 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.879576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.879647 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.879668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.879694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.879713 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:18Z","lastTransitionTime":"2025-12-08T19:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.894011 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:18 crc kubenswrapper[4706]: E1208 19:22:18.894140 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.896099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.896146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.896160 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.896183 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:18 crc kubenswrapper[4706]: I1208 19:22:18.896198 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:18Z","lastTransitionTime":"2025-12-08T19:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:18.999979 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.000017 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.000026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.000040 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.000050 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.103290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.103328 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.103338 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.103353 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.103364 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.206992 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.207035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.207048 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.207068 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.207080 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.310223 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.310344 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.310363 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.310396 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.310415 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.413437 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.413495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.413511 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.413534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.413552 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.516849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.516932 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.516956 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.516984 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.517003 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.608073 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:19 crc kubenswrapper[4706]: E1208 19:22:19.608342 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.620365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.620435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.620453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.620480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.620501 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.723528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.723641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.723669 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.723705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.723740 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.826822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.826888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.826908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.826934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.826951 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.929381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.929463 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.929488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.929523 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:19 crc kubenswrapper[4706]: I1208 19:22:19.929545 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:19Z","lastTransitionTime":"2025-12-08T19:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.032147 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.032203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.032215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.032235 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.032248 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.136400 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.136479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.136511 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.136544 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.136568 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.239781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.239867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.239902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.239935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.239965 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.342882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.343003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.343024 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.343049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.343067 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.352469 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.352631 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.352688 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:22:24.352657828 +0000 UTC m=+26.994858841 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.352737 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.352784 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.352824 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.352855 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.352886 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.352944 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.352950 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.352966 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.352980 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.352995 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.353052 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:24.353033878 +0000 UTC m=+26.995234921 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.353145 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:24.35312975 +0000 UTC m=+26.995330793 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.353191 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:24.353179912 +0000 UTC m=+26.995380955 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.353183 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.353375 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:24.353336246 +0000 UTC m=+26.995537289 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.446819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.446870 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.446884 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.446930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.446948 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.550628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.550676 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.550686 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.550704 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.550717 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.607860 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.607888 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.608054 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:20 crc kubenswrapper[4706]: E1208 19:22:20.608137 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.654013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.654067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.654083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.654102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.654115 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.667174 4706 csr.go:261] certificate signing request csr-p2l9q is approved, waiting to be issued Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.691722 4706 csr.go:257] certificate signing request csr-p2l9q is issued Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.756943 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.756982 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.756991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.757007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.757017 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.795590 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-4546f"] Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.795940 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-4546f" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.801915 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.802577 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.802766 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.830512 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.846489 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.859994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.860050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.860060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.860078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.860091 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.861978 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.862973 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.885364 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.906673 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.929392 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.943079 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.958228 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm22h\" (UniqueName: \"kubernetes.io/projected/7721914f-1dfe-4ddc-a0d8-3b9b42de6440-kube-api-access-fm22h\") pod \"node-resolver-4546f\" (UID: \"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\") " pod="openshift-dns/node-resolver-4546f" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.958315 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7721914f-1dfe-4ddc-a0d8-3b9b42de6440-hosts-file\") pod \"node-resolver-4546f\" (UID: \"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\") " pod="openshift-dns/node-resolver-4546f" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.960107 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.962541 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.962647 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.962670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.962695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.962712 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:20Z","lastTransitionTime":"2025-12-08T19:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.977751 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:20 crc kubenswrapper[4706]: I1208 19:22:20.994702 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:20Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.014409 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.028546 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.041434 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.059521 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7721914f-1dfe-4ddc-a0d8-3b9b42de6440-hosts-file\") pod \"node-resolver-4546f\" (UID: \"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\") " pod="openshift-dns/node-resolver-4546f" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.059939 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm22h\" (UniqueName: \"kubernetes.io/projected/7721914f-1dfe-4ddc-a0d8-3b9b42de6440-kube-api-access-fm22h\") pod \"node-resolver-4546f\" (UID: \"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\") " pod="openshift-dns/node-resolver-4546f" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.059699 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7721914f-1dfe-4ddc-a0d8-3b9b42de6440-hosts-file\") pod \"node-resolver-4546f\" (UID: \"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\") " pod="openshift-dns/node-resolver-4546f" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.060345 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.065291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.065624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.065716 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.065809 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.065939 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.077302 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.082051 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm22h\" (UniqueName: \"kubernetes.io/projected/7721914f-1dfe-4ddc-a0d8-3b9b42de6440-kube-api-access-fm22h\") pod \"node-resolver-4546f\" (UID: \"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\") " pod="openshift-dns/node-resolver-4546f" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.101961 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.114560 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-4546f" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.119747 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: W1208 19:22:21.128595 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7721914f_1dfe_4ddc_a0d8_3b9b42de6440.slice/crio-35da38e11d702dc22daaf0b7c42d8f745a0329437789a2586696ef990876a5f0 WatchSource:0}: Error finding container 35da38e11d702dc22daaf0b7c42d8f745a0329437789a2586696ef990876a5f0: Status 404 returned error can't find the container with id 35da38e11d702dc22daaf0b7c42d8f745a0329437789a2586696ef990876a5f0 Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.131497 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.160194 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.168758 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.168805 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.168819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.168839 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.168857 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.174446 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.206411 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-kp47b"] Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.207009 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.209633 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.210223 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.213707 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.213838 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.216040 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.227024 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.245874 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.271147 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.271186 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.271197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.271214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.271224 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.275915 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.300475 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.332857 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.358893 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.362302 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8be6c92a-aa1d-4749-963b-49e360f63b97-mcd-auth-proxy-config\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.362346 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8be6c92a-aa1d-4749-963b-49e360f63b97-rootfs\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.362372 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8be6c92a-aa1d-4749-963b-49e360f63b97-proxy-tls\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.362409 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njc8q\" (UniqueName: \"kubernetes.io/projected/8be6c92a-aa1d-4749-963b-49e360f63b97-kube-api-access-njc8q\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.373555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.373593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.373603 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.373618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.373627 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.376928 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.393769 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.407011 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.426947 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.443319 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.463626 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8be6c92a-aa1d-4749-963b-49e360f63b97-mcd-auth-proxy-config\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.463660 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8be6c92a-aa1d-4749-963b-49e360f63b97-proxy-tls\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.463685 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8be6c92a-aa1d-4749-963b-49e360f63b97-rootfs\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.463705 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njc8q\" (UniqueName: \"kubernetes.io/projected/8be6c92a-aa1d-4749-963b-49e360f63b97-kube-api-access-njc8q\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.464030 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8be6c92a-aa1d-4749-963b-49e360f63b97-rootfs\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.464849 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8be6c92a-aa1d-4749-963b-49e360f63b97-mcd-auth-proxy-config\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.468678 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8be6c92a-aa1d-4749-963b-49e360f63b97-proxy-tls\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.476069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.476320 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.476380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.476449 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.476505 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.480645 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njc8q\" (UniqueName: \"kubernetes.io/projected/8be6c92a-aa1d-4749-963b-49e360f63b97-kube-api-access-njc8q\") pod \"machine-config-daemon-kp47b\" (UID: \"8be6c92a-aa1d-4749-963b-49e360f63b97\") " pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.522930 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.578871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.578913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.578924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.578942 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.578953 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.609428 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:21 crc kubenswrapper[4706]: E1208 19:22:21.609810 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.628437 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ll6w2"] Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.629212 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-x674x"] Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.629459 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.629540 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.630761 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kl95h"] Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.631904 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.632035 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.632416 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.633481 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.634322 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.634546 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 08 19:22:21 crc kubenswrapper[4706]: W1208 19:22:21.634999 4706 reflector.go:561] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 08 19:22:21 crc kubenswrapper[4706]: E1208 19:22:21.635038 4706 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 08 19:22:21 crc kubenswrapper[4706]: W1208 19:22:21.635045 4706 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 08 19:22:21 crc kubenswrapper[4706]: E1208 19:22:21.635075 4706 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.635138 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.635236 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.635351 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.635402 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.635429 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.635507 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.635668 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.655853 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.675770 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.681220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.681291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.681304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.681324 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.681340 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.689494 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.692653 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-08 19:17:20 +0000 UTC, rotation deadline is 2026-09-19 13:20:57.280830413 +0000 UTC Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.692726 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6833h58m35.58810783s for next certificate rotation Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.707152 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.718150 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.737497 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.757896 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.766453 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-bin\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.766672 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwsjv\" (UniqueName: \"kubernetes.io/projected/1e2bc778-65bf-4304-a4e2-1e31410614c4-kube-api-access-gwsjv\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.766750 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-hostroot\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.766834 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a113dccf-6146-481e-947f-e782eb9c7b53-ovn-node-metrics-cert\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.766927 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-daemon-config\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767016 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9vhk\" (UniqueName: \"kubernetes.io/projected/31d41aa9-3a74-42f4-9933-3f484a89c41d-kube-api-access-q9vhk\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767083 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e2bc778-65bf-4304-a4e2-1e31410614c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767185 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-etc-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767288 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767391 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-cnibin\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767512 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-ovn\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767605 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-os-release\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767692 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-cni-bin\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767766 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-etc-kubernetes\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767840 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767910 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1e2bc778-65bf-4304-a4e2-1e31410614c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.767983 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31d41aa9-3a74-42f4-9933-3f484a89c41d-cni-binary-copy\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768053 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-cnibin\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768133 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-systemd-units\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768243 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-var-lib-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768359 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-system-cni-dir\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768431 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-ovn-kubernetes\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768500 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-netd\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768571 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-k8s-cni-cncf-io\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768656 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-log-socket\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768728 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-socket-dir-parent\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768799 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-conf-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768880 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-netns\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.768954 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-env-overrides\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769028 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-os-release\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769105 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-cni-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769220 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-multus-certs\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769325 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-slash\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769393 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-config\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769473 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-kubelet\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769544 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-script-lib\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769617 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-cni-multus\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769686 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-node-log\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769756 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bptfb\" (UniqueName: \"kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769837 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-kubelet\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769909 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-systemd\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.769977 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-netns\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.770046 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.770126 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-system-cni-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.774778 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.783761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.783917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.784020 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.784109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.784168 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.789630 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.817868 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.866844 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.866901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.866912 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"22ca1c385cdbc4d682662ebca2a6b06653be1b2d0ab7e8336079461b6918a9db"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.868095 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-4546f" event={"ID":"7721914f-1dfe-4ddc-a0d8-3b9b42de6440","Type":"ContainerStarted","Data":"44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.868158 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-4546f" event={"ID":"7721914f-1dfe-4ddc-a0d8-3b9b42de6440","Type":"ContainerStarted","Data":"35da38e11d702dc22daaf0b7c42d8f745a0329437789a2586696ef990876a5f0"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.870933 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-var-lib-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.870973 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-system-cni-dir\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.870996 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-cnibin\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871017 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-systemd-units\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871039 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-ovn-kubernetes\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871059 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-netd\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871102 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-log-socket\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871122 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-socket-dir-parent\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871142 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-k8s-cni-cncf-io\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871162 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-env-overrides\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871183 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-os-release\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871202 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-conf-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871232 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-netns\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871253 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-slash\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871292 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-config\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871318 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-cni-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871342 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-multus-certs\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871367 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-script-lib\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871390 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-cni-multus\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871413 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-kubelet\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871435 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-node-log\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871460 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bptfb\" (UniqueName: \"kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871484 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-systemd\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871507 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-netns\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871529 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-kubelet\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871550 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871572 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-system-cni-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871602 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-bin\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871639 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwsjv\" (UniqueName: \"kubernetes.io/projected/1e2bc778-65bf-4304-a4e2-1e31410614c4-kube-api-access-gwsjv\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871660 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-hostroot\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871680 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9vhk\" (UniqueName: \"kubernetes.io/projected/31d41aa9-3a74-42f4-9933-3f484a89c41d-kube-api-access-q9vhk\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871708 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a113dccf-6146-481e-947f-e782eb9c7b53-ovn-node-metrics-cert\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-daemon-config\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871748 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871766 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-cnibin\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871781 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e2bc778-65bf-4304-a4e2-1e31410614c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871797 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-etc-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871812 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-cni-bin\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871829 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-etc-kubernetes\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871848 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-ovn\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871867 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-os-release\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871886 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1e2bc778-65bf-4304-a4e2-1e31410614c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871905 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31d41aa9-3a74-42f4-9933-3f484a89c41d-cni-binary-copy\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.871925 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872177 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-var-lib-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872248 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-system-cni-dir\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872350 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-cnibin\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872376 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-systemd-units\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872372 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872413 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-kubelet\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872421 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-netns\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872437 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872469 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-ovn-kubernetes\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872495 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-netd\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872519 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-log-socket\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872559 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-socket-dir-parent\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872565 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-system-cni-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872591 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-k8s-cni-cncf-io\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872594 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-bin\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.872871 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-hostroot\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873008 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-cni-bin\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873087 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-kubelet\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873109 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873178 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-run-multus-certs\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873201 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-cnibin\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873280 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-os-release\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873331 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-slash\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873325 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-netns\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873393 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-ovn\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873397 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e2bc778-65bf-4304-a4e2-1e31410614c4-os-release\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873451 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-conf-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873514 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-node-log\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873589 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-env-overrides\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873628 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-systemd\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873635 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-cni-dir\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873825 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-host-var-lib-cni-multus\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.873938 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-script-lib\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.874001 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31d41aa9-3a74-42f4-9933-3f484a89c41d-etc-kubernetes\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.874080 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-etc-openvswitch\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.874092 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1e2bc778-65bf-4304-a4e2-1e31410614c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.874248 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e2bc778-65bf-4304-a4e2-1e31410614c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.874254 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-config\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.874500 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31d41aa9-3a74-42f4-9933-3f484a89c41d-multus-daemon-config\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.874511 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31d41aa9-3a74-42f4-9933-3f484a89c41d-cni-binary-copy\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.878667 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.878985 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a113dccf-6146-481e-947f-e782eb9c7b53-ovn-node-metrics-cert\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.889580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.889630 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.889643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.889662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.889674 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.904571 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.904991 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwsjv\" (UniqueName: \"kubernetes.io/projected/1e2bc778-65bf-4304-a4e2-1e31410614c4-kube-api-access-gwsjv\") pod \"multus-additional-cni-plugins-ll6w2\" (UID: \"1e2bc778-65bf-4304-a4e2-1e31410614c4\") " pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.925687 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9vhk\" (UniqueName: \"kubernetes.io/projected/31d41aa9-3a74-42f4-9933-3f484a89c41d-kube-api-access-q9vhk\") pod \"multus-x674x\" (UID: \"31d41aa9-3a74-42f4-9933-3f484a89c41d\") " pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.939056 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.942909 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.952969 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x674x" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.971290 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:21 crc kubenswrapper[4706]: W1208 19:22:21.975753 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31d41aa9_3a74_42f4_9933_3f484a89c41d.slice/crio-067fb83efaaf82e2211f0810b841d9fca4a2e4b353997c95f814e29d0a358e3b WatchSource:0}: Error finding container 067fb83efaaf82e2211f0810b841d9fca4a2e4b353997c95f814e29d0a358e3b: Status 404 returned error can't find the container with id 067fb83efaaf82e2211f0810b841d9fca4a2e4b353997c95f814e29d0a358e3b Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.992389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.992439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.992453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.992472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.992484 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:21Z","lastTransitionTime":"2025-12-08T19:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:21 crc kubenswrapper[4706]: I1208 19:22:21.993641 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:21Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.016744 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.031850 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.050666 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.064140 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.096944 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.097000 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.097025 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.097051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.097068 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.102460 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.116621 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.132434 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.149344 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.162223 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.175396 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.192209 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.203682 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.203906 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.203970 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.204041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.204129 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.208054 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.308075 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.308116 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.308127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.308145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.308158 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.410743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.410778 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.410788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.410804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.410815 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.513230 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.513302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.513316 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.513338 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.513352 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.607338 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.607337 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:22 crc kubenswrapper[4706]: E1208 19:22:22.607796 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:22 crc kubenswrapper[4706]: E1208 19:22:22.607797 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.616927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.617381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.617397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.617418 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.617479 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.673089 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.720119 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.720171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.720183 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.720203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.720217 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.823258 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.823329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.823343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.823366 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.823382 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.872542 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x674x" event={"ID":"31d41aa9-3a74-42f4-9933-3f484a89c41d","Type":"ContainerStarted","Data":"92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.872601 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x674x" event={"ID":"31d41aa9-3a74-42f4-9933-3f484a89c41d","Type":"ContainerStarted","Data":"067fb83efaaf82e2211f0810b841d9fca4a2e4b353997c95f814e29d0a358e3b"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.875311 4706 generic.go:334] "Generic (PLEG): container finished" podID="1e2bc778-65bf-4304-a4e2-1e31410614c4" containerID="bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542" exitCode=0 Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.875362 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" event={"ID":"1e2bc778-65bf-4304-a4e2-1e31410614c4","Type":"ContainerDied","Data":"bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.875391 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" event={"ID":"1e2bc778-65bf-4304-a4e2-1e31410614c4","Type":"ContainerStarted","Data":"4ccd37952956c610673c06b44e17e61e1bdc7ed71212e27960d1812455f563ac"} Dec 08 19:22:22 crc kubenswrapper[4706]: E1208 19:22:22.890736 4706 projected.go:288] Couldn't get configMap openshift-ovn-kubernetes/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 08 19:22:22 crc kubenswrapper[4706]: E1208 19:22:22.890782 4706 projected.go:194] Error preparing data for projected volume kube-api-access-bptfb for pod openshift-ovn-kubernetes/ovnkube-node-kl95h: failed to sync configmap cache: timed out waiting for the condition Dec 08 19:22:22 crc kubenswrapper[4706]: E1208 19:22:22.890874 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb podName:a113dccf-6146-481e-947f-e782eb9c7b53 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:23.390845523 +0000 UTC m=+26.033046526 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-bptfb" (UniqueName: "kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb") pod "ovnkube-node-kl95h" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53") : failed to sync configmap cache: timed out waiting for the condition Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.898681 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.920141 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.925617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.925662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.925673 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.925695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.925707 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:22Z","lastTransitionTime":"2025-12-08T19:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.933867 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.946019 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.960874 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.973946 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:22 crc kubenswrapper[4706]: I1208 19:22:22.986784 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.001741 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:22Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.013555 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.027356 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.028767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.028797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.028806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.028824 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.028837 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.042320 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.053964 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.067115 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.085470 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.097784 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.108306 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.115021 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-qxrgm"] Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.115395 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.117430 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.117430 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.117487 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.117949 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.123905 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.134109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.134170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.134183 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.134202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.134219 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.144963 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.160236 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.173943 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.187405 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.191190 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.203995 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.221803 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.236972 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.237019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.237031 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.237054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.237067 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.237755 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.253521 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.273151 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.285586 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.287053 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f0cd396c-f997-4a5c-b5ec-a901d099bc63-serviceca\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.287140 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0cd396c-f997-4a5c-b5ec-a901d099bc63-host\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.287165 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpnbq\" (UniqueName: \"kubernetes.io/projected/f0cd396c-f997-4a5c-b5ec-a901d099bc63-kube-api-access-bpnbq\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.300616 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.313209 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.327644 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.341050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.341424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.341436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.341455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.341468 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.343522 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.356817 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.375675 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.388617 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0cd396c-f997-4a5c-b5ec-a901d099bc63-host\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.388674 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpnbq\" (UniqueName: \"kubernetes.io/projected/f0cd396c-f997-4a5c-b5ec-a901d099bc63-kube-api-access-bpnbq\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.388726 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f0cd396c-f997-4a5c-b5ec-a901d099bc63-serviceca\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.388783 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0cd396c-f997-4a5c-b5ec-a901d099bc63-host\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.389712 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f0cd396c-f997-4a5c-b5ec-a901d099bc63-serviceca\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.394190 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.406959 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.409785 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpnbq\" (UniqueName: \"kubernetes.io/projected/f0cd396c-f997-4a5c-b5ec-a901d099bc63-kube-api-access-bpnbq\") pod \"node-ca-qxrgm\" (UID: \"f0cd396c-f997-4a5c-b5ec-a901d099bc63\") " pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.418057 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.429116 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.432379 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qxrgm" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.441852 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.443780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.443820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.443831 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.443853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.443865 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: W1208 19:22:23.446912 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0cd396c_f997_4a5c_b5ec_a901d099bc63.slice/crio-09a2d62db343b8129d268945931c8760db64f7a71e4a808fa9904a85158d1a95 WatchSource:0}: Error finding container 09a2d62db343b8129d268945931c8760db64f7a71e4a808fa9904a85158d1a95: Status 404 returned error can't find the container with id 09a2d62db343b8129d268945931c8760db64f7a71e4a808fa9904a85158d1a95 Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.455782 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.468477 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.480920 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.490195 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bptfb\" (UniqueName: \"kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.495413 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bptfb\" (UniqueName: \"kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb\") pod \"ovnkube-node-kl95h\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.495536 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.530876 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.546097 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.546138 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.546148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.546162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.546173 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.608038 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:23 crc kubenswrapper[4706]: E1208 19:22:23.608211 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.647871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.647901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.647908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.647923 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.647933 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.750997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.751046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.751055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.751073 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.751084 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.762396 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:23 crc kubenswrapper[4706]: W1208 19:22:23.775329 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda113dccf_6146_481e_947f_e782eb9c7b53.slice/crio-6777c21a19e6519123cc4ee0a68eaa48655bfc2dc2ad40cca941fbc6ae39ad6c WatchSource:0}: Error finding container 6777c21a19e6519123cc4ee0a68eaa48655bfc2dc2ad40cca941fbc6ae39ad6c: Status 404 returned error can't find the container with id 6777c21a19e6519123cc4ee0a68eaa48655bfc2dc2ad40cca941fbc6ae39ad6c Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.853832 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.853875 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.853887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.853938 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.853953 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.883048 4706 generic.go:334] "Generic (PLEG): container finished" podID="1e2bc778-65bf-4304-a4e2-1e31410614c4" containerID="08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0" exitCode=0 Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.883211 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" event={"ID":"1e2bc778-65bf-4304-a4e2-1e31410614c4","Type":"ContainerDied","Data":"08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.886121 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944" exitCode=0 Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.886206 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.886242 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"6777c21a19e6519123cc4ee0a68eaa48655bfc2dc2ad40cca941fbc6ae39ad6c"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.889994 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qxrgm" event={"ID":"f0cd396c-f997-4a5c-b5ec-a901d099bc63","Type":"ContainerStarted","Data":"fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.890058 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qxrgm" event={"ID":"f0cd396c-f997-4a5c-b5ec-a901d099bc63","Type":"ContainerStarted","Data":"09a2d62db343b8129d268945931c8760db64f7a71e4a808fa9904a85158d1a95"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.898500 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.914522 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.928567 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.942013 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.955030 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.957008 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.957032 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.957060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.957076 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.957086 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:23Z","lastTransitionTime":"2025-12-08T19:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.969105 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:23 crc kubenswrapper[4706]: I1208 19:22:23.988402 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:23Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.005224 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.019732 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.034734 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.055309 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.060977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.061027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.061074 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.061093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.061106 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.079869 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.092294 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.106089 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.124610 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.163012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.163102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.163131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.163165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.163194 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.168107 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.208355 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.251986 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.267019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.267067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.267079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.267099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.267115 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.296214 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.329714 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.369478 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.370209 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.370280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.370296 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.370317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.370331 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.404672 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.404785 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.404815 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.404834 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.404868 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.404949 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:22:32.404927356 +0000 UTC m=+35.047128359 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.404992 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405057 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405063 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405082 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405086 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405088 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:32.40506883 +0000 UTC m=+35.047269833 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405102 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405104 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405002 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405152 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:32.405137582 +0000 UTC m=+35.047338595 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405193 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:32.405183433 +0000 UTC m=+35.047384456 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.405211 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:32.405202373 +0000 UTC m=+35.047403386 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.408643 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.445814 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.472787 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.472834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.472847 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.472867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.472879 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.497895 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.530144 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.569086 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.575460 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.575518 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.575531 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.575550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.575563 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.607608 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.607608 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.607794 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:24 crc kubenswrapper[4706]: E1208 19:22:24.607826 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.608403 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.645222 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.677680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.677719 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.677734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.677752 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.677765 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.689539 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.728872 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.781112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.781167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.781180 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.781197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.781213 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.884934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.885018 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.885043 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.885079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.885105 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.899319 4706 generic.go:334] "Generic (PLEG): container finished" podID="1e2bc778-65bf-4304-a4e2-1e31410614c4" containerID="946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754" exitCode=0 Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.899337 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" event={"ID":"1e2bc778-65bf-4304-a4e2-1e31410614c4","Type":"ContainerDied","Data":"946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.906610 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.906683 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.906707 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.906723 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.906736 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.906749 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.925643 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.940114 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.960145 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.978453 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.987960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.988012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.988025 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.988041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.988053 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:24Z","lastTransitionTime":"2025-12-08T19:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:24 crc kubenswrapper[4706]: I1208 19:22:24.996207 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:24Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.018703 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.038846 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.063056 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.089780 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.093961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.094011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.094023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.094045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.094061 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.129054 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.169729 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.196964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.196999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.197009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.197029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.197040 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.206808 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.260620 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.293913 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.300555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.300621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.300634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.300655 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.300670 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.330965 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.403627 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.403695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.403712 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.403739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.403769 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.506597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.506661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.506680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.506707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.506728 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.607531 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:25 crc kubenswrapper[4706]: E1208 19:22:25.607793 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.610356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.610427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.610438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.610459 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.610470 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.713488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.713600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.713628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.713665 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.713695 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.817171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.817224 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.817235 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.817252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.817284 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.913856 4706 generic.go:334] "Generic (PLEG): container finished" podID="1e2bc778-65bf-4304-a4e2-1e31410614c4" containerID="8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776" exitCode=0 Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.913929 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" event={"ID":"1e2bc778-65bf-4304-a4e2-1e31410614c4","Type":"ContainerDied","Data":"8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.918995 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.919040 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.919051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.919079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.919090 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:25Z","lastTransitionTime":"2025-12-08T19:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.928789 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.946985 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.962893 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.979001 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:25 crc kubenswrapper[4706]: I1208 19:22:25.996252 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:25Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.011554 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.021845 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.021902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.021919 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.021946 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.021961 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.030021 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.045142 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.059673 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.080798 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.099562 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.115216 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.124925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.124985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.124999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.125018 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.125035 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.130027 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.139193 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.161472 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.228255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.228666 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.228825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.228963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.229185 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.332060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.332099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.332108 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.332124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.332133 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.440389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.440452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.440466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.440487 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.440512 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.544240 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.544805 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.544833 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.544865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.544905 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.607948 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.608030 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:26 crc kubenswrapper[4706]: E1208 19:22:26.608469 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:26 crc kubenswrapper[4706]: E1208 19:22:26.608635 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.647598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.647641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.647653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.647672 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.647686 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.751203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.751308 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.751323 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.751348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.751377 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.855305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.855392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.855413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.855433 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.855450 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.921874 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.925658 4706 generic.go:334] "Generic (PLEG): container finished" podID="1e2bc778-65bf-4304-a4e2-1e31410614c4" containerID="e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663" exitCode=0 Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.925734 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" event={"ID":"1e2bc778-65bf-4304-a4e2-1e31410614c4","Type":"ContainerDied","Data":"e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.942751 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.956103 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.958001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.958054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.958096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.958119 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.958180 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:26Z","lastTransitionTime":"2025-12-08T19:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.969337 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.980364 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:26 crc kubenswrapper[4706]: I1208 19:22:26.993117 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:26Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.004883 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.016664 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.029734 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.048805 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.060650 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.060694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.060710 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.060735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.060977 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.062869 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.081538 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.094430 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.104392 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.123560 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.139527 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.164131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.164195 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.164209 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.164241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.164259 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.267730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.267776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.267787 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.267804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.267814 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.371090 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.371133 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.371143 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.371163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.371174 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.477235 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.477316 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.477330 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.477352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.477366 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.581343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.581408 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.581427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.581455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.581476 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.609915 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:27 crc kubenswrapper[4706]: E1208 19:22:27.610034 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.628494 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.646402 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.657324 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.672406 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.684702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.684761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.684774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.684813 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.684826 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.690095 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.705917 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.720244 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.734965 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.746524 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.765414 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.787635 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.787688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.787699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.787721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.787734 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.788212 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.791760 4706 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.819335 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.837375 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.852107 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.889977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.890014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.890057 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.890078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.890092 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.897153 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.934063 4706 generic.go:334] "Generic (PLEG): container finished" podID="1e2bc778-65bf-4304-a4e2-1e31410614c4" containerID="a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f" exitCode=0 Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.934152 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" event={"ID":"1e2bc778-65bf-4304-a4e2-1e31410614c4","Type":"ContainerDied","Data":"a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f"} Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.957434 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:27 crc kubenswrapper[4706]: I1208 19:22:27.979807 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:27Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:27.994211 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:27.994253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:27.994277 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:27.994293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:27.994304 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:27Z","lastTransitionTime":"2025-12-08T19:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.010608 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.026456 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.040588 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.053786 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.066624 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.080774 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.096045 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.096550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.096584 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.096598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.096617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.096630 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.109473 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.126409 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.143204 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.165561 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.184137 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.199691 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.200045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.200168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.200293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.200403 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.203410 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.303287 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.303343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.303357 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.303380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.303394 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.406919 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.406971 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.406983 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.407004 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.407018 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.510232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.510333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.510356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.510385 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.510403 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.608314 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.608322 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:28 crc kubenswrapper[4706]: E1208 19:22:28.608563 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:28 crc kubenswrapper[4706]: E1208 19:22:28.608709 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.650884 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.650933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.650943 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.650962 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.650975 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.754568 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.754615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.754628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.754648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.754660 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.858016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.858068 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.858079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.858100 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.858113 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.918488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.918562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.918580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.918610 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.918630 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: E1208 19:22:28.938787 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.942658 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" event={"ID":"1e2bc778-65bf-4304-a4e2-1e31410614c4","Type":"ContainerStarted","Data":"4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.943748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.943795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.943808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.943826 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.943839 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: E1208 19:22:28.959517 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.965555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.965616 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.965633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.965656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.965672 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.966474 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: E1208 19:22:28.982900 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.992881 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:28Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.994043 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.994069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.994081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.994105 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:28 crc kubenswrapper[4706]: I1208 19:22:28.994122 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:28Z","lastTransitionTime":"2025-12-08T19:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: E1208 19:22:29.007898 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.011201 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.014017 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.014053 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.014064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.014083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.014095 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.024434 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: E1208 19:22:29.029250 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: E1208 19:22:29.029674 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.032853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.032914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.032936 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.032977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.032999 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.039721 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.056847 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.071211 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.088346 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.107891 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.120346 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.135714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.135753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.135763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.135779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.135791 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.147700 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.174845 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.193137 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.209203 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.223993 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.238109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.238462 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.238559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.238647 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.238724 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.341995 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.342039 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.342052 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.342074 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.342086 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.444758 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.444808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.444818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.444833 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.444842 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.548901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.548994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.549019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.549055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.549080 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.608245 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:29 crc kubenswrapper[4706]: E1208 19:22:29.608526 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.652601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.652673 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.652692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.652721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.652741 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.756721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.756786 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.756796 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.756811 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.756822 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.860166 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.860243 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.860301 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.860335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.860358 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.963452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.963507 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.963521 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.963542 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.963556 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:29Z","lastTransitionTime":"2025-12-08T19:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.969566 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41"} Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.970054 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:29 crc kubenswrapper[4706]: I1208 19:22:29.990989 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:29Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:29.999993 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.013097 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.030967 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.046288 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.061940 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.065970 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.066013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.066029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.066054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.066073 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.075637 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.097531 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.117768 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.133043 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.148234 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.161989 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.168964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.169006 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.169017 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.169036 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.169051 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.174518 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.188525 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.203813 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.219388 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.232878 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.250767 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.269703 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.271189 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.271252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.271282 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.271300 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.271317 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.288369 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.305089 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.320805 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.342525 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.364155 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.374525 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.374590 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.374606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.374638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.374656 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.388952 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.405244 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.433084 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.462515 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.477744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.477803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.477819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.477844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.477863 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.492239 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.509469 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.523078 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.581095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.581467 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.581534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.581606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.581672 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.607316 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.607406 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:30 crc kubenswrapper[4706]: E1208 19:22:30.607489 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:30 crc kubenswrapper[4706]: E1208 19:22:30.607582 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.684840 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.685251 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.685346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.685466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.685548 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.788853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.788902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.788912 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.788930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.788944 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.893023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.893669 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.893692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.893722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.893742 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.952515 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.973147 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.973738 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.976840 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:30Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.996723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.996784 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.996800 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.996825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:30 crc kubenswrapper[4706]: I1208 19:22:30.996842 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:30Z","lastTransitionTime":"2025-12-08T19:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.002761 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.014150 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.020254 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.035844 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.048599 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.060782 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.075399 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.092030 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.099522 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.099559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.099569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.099587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.099598 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.105788 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.120860 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.144330 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.169008 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.183156 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.198465 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.203011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.203068 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.203083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.203106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.203118 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.215118 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.231849 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.251590 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.268807 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.288531 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.305762 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.306332 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.306398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.306414 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.306439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.306453 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.318784 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.336546 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.354602 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.372063 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.387990 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.406624 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.412701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.412740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.412749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.412765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.412775 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.430235 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.442157 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.455296 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.465256 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.515946 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.515993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.516007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.516027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.516041 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.607901 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:31 crc kubenswrapper[4706]: E1208 19:22:31.608066 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.618733 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.618774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.618782 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.618798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.618808 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.721549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.721602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.721610 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.721628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.721639 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.824582 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.824631 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.824643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.824658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.824668 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.927154 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.927190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.927199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.927214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.927223 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:31Z","lastTransitionTime":"2025-12-08T19:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.977663 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/0.log" Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.980497 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41" exitCode=1 Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.980554 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41"} Dec 08 19:22:31 crc kubenswrapper[4706]: I1208 19:22:31.981180 4706 scope.go:117] "RemoveContainer" containerID="b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.001313 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:31Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.013076 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.025666 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.029494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.029546 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.029559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.029576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.029589 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.037199 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.047511 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.057939 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.070350 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.083510 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.095685 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.108483 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.126256 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:31Z\\\",\\\"message\\\":\\\":22:31.141636 6017 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 19:22:31.141672 6017 factory.go:656] Stopping watch factory\\\\nI1208 19:22:31.141689 6017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 19:22:31.141722 6017 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.141908 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 19:22:31.142086 6017 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142167 6017 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 19:22:31.142368 6017 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142575 6017 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142640 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.132172 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.132210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.132219 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.132233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.132242 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.139697 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.151623 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.162424 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.174199 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.234357 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.234394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.234406 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.234426 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.234438 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.336751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.336792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.336802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.336817 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.336830 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.439777 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.439821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.439833 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.439854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.439865 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.486993 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.487108 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.487657 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:22:48.487148502 +0000 UTC m=+51.129349505 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.487759 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.487787 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.487820 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.487895 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:48.487862722 +0000 UTC m=+51.130063725 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.491358 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.491515 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.491550 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.491610 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.491644 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:48.491613694 +0000 UTC m=+51.133814697 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.491708 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.491765 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:48.491752008 +0000 UTC m=+51.133953011 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.491809 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.491827 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.491841 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.491878 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 19:22:48.491870211 +0000 UTC m=+51.134071214 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.542454 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.542506 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.542517 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.542537 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.542551 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.608029 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.608050 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.608293 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:32 crc kubenswrapper[4706]: E1208 19:22:32.608346 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.645198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.645254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.645291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.645309 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.645321 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.748387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.748456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.748476 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.748501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.748516 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.851895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.851946 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.851963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.851987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.852005 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.954354 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.954405 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.954417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.954436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.954449 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:32Z","lastTransitionTime":"2025-12-08T19:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.985530 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/0.log" Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.989490 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91"} Dec 08 19:22:32 crc kubenswrapper[4706]: I1208 19:22:32.989662 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.006352 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.045094 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:31Z\\\",\\\"message\\\":\\\":22:31.141636 6017 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 19:22:31.141672 6017 factory.go:656] Stopping watch factory\\\\nI1208 19:22:31.141689 6017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 19:22:31.141722 6017 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.141908 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 19:22:31.142086 6017 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142167 6017 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 19:22:31.142368 6017 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142575 6017 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142640 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.057700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.057738 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.057751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.057770 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.057782 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.076104 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.098504 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.116880 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.128188 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.167973 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.168041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.168060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.168091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.168115 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.174972 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.195061 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.216156 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.236620 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.250157 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.265775 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.271040 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.271092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.271105 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.271126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.271141 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.286964 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.307330 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.325563 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:33Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.373598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.373681 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.373705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.373734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.373755 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.476289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.476617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.476690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.476772 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.476857 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.580929 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.580986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.581008 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.581038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.581065 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.608203 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:33 crc kubenswrapper[4706]: E1208 19:22:33.608462 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.685515 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.686026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.686221 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.686426 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.686579 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.789825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.790325 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.790348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.790376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.790399 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.900783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.900835 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.900852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.900871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.900883 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:33Z","lastTransitionTime":"2025-12-08T19:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.997232 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/1.log" Dec 08 19:22:33 crc kubenswrapper[4706]: I1208 19:22:33.998563 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/0.log" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.004317 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91" exitCode=1 Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.004385 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.004473 4706 scope.go:117] "RemoveContainer" containerID="b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.005984 4706 scope.go:117] "RemoveContainer" containerID="1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91" Dec 08 19:22:34 crc kubenswrapper[4706]: E1208 19:22:34.006172 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.007721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.007749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.007757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.007795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.007809 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.024248 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b"] Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.024519 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.025029 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.027712 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.028608 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.043422 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.058289 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.072603 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.085731 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.096974 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.109468 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.110414 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.110514 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpkpg\" (UniqueName: \"kubernetes.io/projected/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-kube-api-access-lpkpg\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.110620 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.111602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.111643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.111656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.111676 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.111690 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.112456 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.125816 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.139187 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.152985 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.180383 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:31Z\\\",\\\"message\\\":\\\":22:31.141636 6017 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 19:22:31.141672 6017 factory.go:656] Stopping watch factory\\\\nI1208 19:22:31.141689 6017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 19:22:31.141722 6017 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.141908 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 19:22:31.142086 6017 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142167 6017 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 19:22:31.142368 6017 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142575 6017 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142640 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.211697 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.211946 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.212047 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.212111 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.212155 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpkpg\" (UniqueName: \"kubernetes.io/projected/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-kube-api-access-lpkpg\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.212942 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.214484 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.215361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.215419 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.215443 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.215479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.215504 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.225159 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.231807 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.235679 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpkpg\" (UniqueName: \"kubernetes.io/projected/3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc-kube-api-access-lpkpg\") pod \"ovnkube-control-plane-749d76644c-sz99b\" (UID: \"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.253016 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.267079 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.287891 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.326761 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.329120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.329172 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.329193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.329217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.329232 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.341125 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.351465 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3a1168cd08174ed63132f8b8307b5776afdcbcb681336311a5940dd2e468c41\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:31Z\\\",\\\"message\\\":\\\":22:31.141636 6017 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 19:22:31.141672 6017 factory.go:656] Stopping watch factory\\\\nI1208 19:22:31.141689 6017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 19:22:31.141722 6017 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.141908 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 19:22:31.142086 6017 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142167 6017 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 19:22:31.142368 6017 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142575 6017 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 19:22:31.142640 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: W1208 19:22:34.357701 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bc33a1a_7cae_4be1_9348_d3f9ebbe84cc.slice/crio-93007be9ecf22f96337b779f7b46445a84c4495738237f2091498226fb3dd8cf WatchSource:0}: Error finding container 93007be9ecf22f96337b779f7b46445a84c4495738237f2091498226fb3dd8cf: Status 404 returned error can't find the container with id 93007be9ecf22f96337b779f7b46445a84c4495738237f2091498226fb3dd8cf Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.366950 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.387299 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.404409 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.415375 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.432312 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.432401 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.432438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.432456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.433360 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.438692 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.451641 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.469461 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.486518 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.500181 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.515544 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.530139 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.537905 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.538162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.538333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.538439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.538541 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.547556 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.562228 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:34Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.607260 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.607308 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:34 crc kubenswrapper[4706]: E1208 19:22:34.607430 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:34 crc kubenswrapper[4706]: E1208 19:22:34.607503 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.641225 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.641290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.641305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.641324 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.641337 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.744342 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.744387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.744397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.744419 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.744435 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.848214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.848260 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.848315 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.848334 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.848346 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.951545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.951591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.951601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.951619 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:34 crc kubenswrapper[4706]: I1208 19:22:34.951636 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:34Z","lastTransitionTime":"2025-12-08T19:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.011990 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/1.log" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.015801 4706 scope.go:117] "RemoveContainer" containerID="1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91" Dec 08 19:22:35 crc kubenswrapper[4706]: E1208 19:22:35.015981 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.018410 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" event={"ID":"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc","Type":"ContainerStarted","Data":"3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.018457 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" event={"ID":"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc","Type":"ContainerStarted","Data":"0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.018470 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" event={"ID":"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc","Type":"ContainerStarted","Data":"93007be9ecf22f96337b779f7b46445a84c4495738237f2091498226fb3dd8cf"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.044451 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.055639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.055730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.055752 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.055784 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.055803 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.073886 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.097577 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.120156 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.158480 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.160246 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.160340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.160360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.160387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.160407 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.198991 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.219996 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.236556 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.251179 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.263595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.263694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.263735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.263774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.263804 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.273527 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.298829 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.321063 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.344129 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.357555 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.366651 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.366718 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.366741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.366774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.366799 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.371234 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.385447 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.398546 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.429505 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.451468 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.469255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.469317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.469326 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.469342 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.469354 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.474368 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.491665 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.510233 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.523164 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.535460 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.538996 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-strs6"] Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.539558 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:35 crc kubenswrapper[4706]: E1208 19:22:35.539621 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.552280 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.566081 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.572010 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.572076 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.572088 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.572126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.572140 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.578666 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.591053 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.608004 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:35 crc kubenswrapper[4706]: E1208 19:22:35.608165 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.609838 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.629944 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.645129 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.645250 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpl97\" (UniqueName: \"kubernetes.io/projected/439dd590-d737-435b-80a7-01f75fe6b64d-kube-api-access-kpl97\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.653763 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.670191 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.676076 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.676157 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.676183 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.676210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.676230 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.684915 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.701699 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.718433 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.732290 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.746662 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpl97\" (UniqueName: \"kubernetes.io/projected/439dd590-d737-435b-80a7-01f75fe6b64d-kube-api-access-kpl97\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.746765 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:35 crc kubenswrapper[4706]: E1208 19:22:35.746903 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:35 crc kubenswrapper[4706]: E1208 19:22:35.746971 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs podName:439dd590-d737-435b-80a7-01f75fe6b64d nodeName:}" failed. No retries permitted until 2025-12-08 19:22:36.24695062 +0000 UTC m=+38.889151623 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs") pod "network-metrics-daemon-strs6" (UID: "439dd590-d737-435b-80a7-01f75fe6b64d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.748600 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.760785 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.764288 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpl97\" (UniqueName: \"kubernetes.io/projected/439dd590-d737-435b-80a7-01f75fe6b64d-kube-api-access-kpl97\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.776214 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.778709 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.778738 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.778747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.778760 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.778770 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.789793 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.803787 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.815568 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.826691 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.846212 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.865901 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.881165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.881440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.881503 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.881540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.881565 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.895745 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.910693 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.924481 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.939404 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:35Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.984574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.984632 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.984648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.984674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:35 crc kubenswrapper[4706]: I1208 19:22:35.984735 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:35Z","lastTransitionTime":"2025-12-08T19:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.088287 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.088345 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.088363 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.088383 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.088396 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.192195 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.192244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.192255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.192305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.192319 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.251819 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:36 crc kubenswrapper[4706]: E1208 19:22:36.252060 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:36 crc kubenswrapper[4706]: E1208 19:22:36.252222 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs podName:439dd590-d737-435b-80a7-01f75fe6b64d nodeName:}" failed. No retries permitted until 2025-12-08 19:22:37.252176425 +0000 UTC m=+39.894377598 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs") pod "network-metrics-daemon-strs6" (UID: "439dd590-d737-435b-80a7-01f75fe6b64d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.302431 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.302494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.302506 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.302523 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.302534 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.406369 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.406419 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.406433 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.406452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.406466 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.509923 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.510170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.510190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.510221 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.510245 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.608392 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.608416 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:36 crc kubenswrapper[4706]: E1208 19:22:36.608706 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:36 crc kubenswrapper[4706]: E1208 19:22:36.608817 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.609251 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:36 crc kubenswrapper[4706]: E1208 19:22:36.609898 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.613749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.613808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.613828 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.613851 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.613873 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.717603 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.717664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.717680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.717706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.717723 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.821654 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.822070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.822219 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.822514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.822754 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.926491 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.926611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.926640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.926673 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:36 crc kubenswrapper[4706]: I1208 19:22:36.926757 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:36Z","lastTransitionTime":"2025-12-08T19:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.018177 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.019707 4706 scope.go:117] "RemoveContainer" containerID="1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91" Dec 08 19:22:37 crc kubenswrapper[4706]: E1208 19:22:37.020006 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.029554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.029615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.029636 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.029659 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.029676 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.133010 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.133069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.133081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.133100 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.133119 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.235903 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.235952 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.235969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.235986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.236001 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.263515 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:37 crc kubenswrapper[4706]: E1208 19:22:37.263794 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:37 crc kubenswrapper[4706]: E1208 19:22:37.263952 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs podName:439dd590-d737-435b-80a7-01f75fe6b64d nodeName:}" failed. No retries permitted until 2025-12-08 19:22:39.263918221 +0000 UTC m=+41.906119414 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs") pod "network-metrics-daemon-strs6" (UID: "439dd590-d737-435b-80a7-01f75fe6b64d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.339305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.339394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.339408 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.339430 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.339469 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.443360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.443712 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.443849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.443964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.444068 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.548112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.548518 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.548630 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.548716 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.548828 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.608031 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:37 crc kubenswrapper[4706]: E1208 19:22:37.608772 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.637953 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.651638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.651693 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.651705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.651723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.651749 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.657108 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.671087 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.687611 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.708756 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.731092 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.748158 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.755067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.755141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.755169 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.755206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.755234 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.765668 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.787091 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.805718 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.826178 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.846039 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.858212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.858298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.858318 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.858345 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.858365 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.861233 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.878107 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.892831 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.906870 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.921726 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:37Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.961275 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.961331 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.961342 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.961358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:37 crc kubenswrapper[4706]: I1208 19:22:37.961369 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:37Z","lastTransitionTime":"2025-12-08T19:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.064365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.064401 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.064412 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.064427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.064440 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.166893 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.166981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.166995 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.167013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.167027 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.270406 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.270504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.270529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.270565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.270592 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.374005 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.374072 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.374091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.374122 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.374146 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.477052 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.477117 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.477126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.477139 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.477148 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.579380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.579426 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.579438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.579456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.579467 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.607929 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.607972 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.607989 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:38 crc kubenswrapper[4706]: E1208 19:22:38.608105 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:38 crc kubenswrapper[4706]: E1208 19:22:38.608196 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:38 crc kubenswrapper[4706]: E1208 19:22:38.608326 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.683045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.683096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.683106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.683123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.683139 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.785624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.785700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.785716 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.785741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.785755 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.888406 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.888486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.888500 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.888518 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.888531 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.991909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.991983 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.992003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.992034 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:38 crc kubenswrapper[4706]: I1208 19:22:38.992061 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:38Z","lastTransitionTime":"2025-12-08T19:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.095087 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.095154 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.095173 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.095195 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.095208 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.197946 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.198001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.198013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.198028 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.198039 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.292166 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.292373 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.292458 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs podName:439dd590-d737-435b-80a7-01f75fe6b64d nodeName:}" failed. No retries permitted until 2025-12-08 19:22:43.292434711 +0000 UTC m=+45.934635714 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs") pod "network-metrics-daemon-strs6" (UID: "439dd590-d737-435b-80a7-01f75fe6b64d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.302098 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.302163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.302188 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.302221 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.302242 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.354981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.355070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.355087 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.355107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.355125 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.373678 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:39Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.379019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.379085 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.379108 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.379145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.379173 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.394104 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:39Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.398754 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.398836 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.398852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.398877 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.398894 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.413221 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:39Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.417852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.417902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.417916 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.417936 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.417952 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.431500 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:39Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.436736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.436787 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.436801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.436822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.436837 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.450166 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:39Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.450446 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.452680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.452734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.452747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.452766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.452778 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.555432 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.555486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.555496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.555515 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.555526 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.608665 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:39 crc kubenswrapper[4706]: E1208 19:22:39.608864 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.658705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.658780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.658802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.658826 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.658842 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.762844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.762897 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.762909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.762928 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.762941 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.865321 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.865387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.865412 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.865468 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.865494 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.968710 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.968755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.968767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.968784 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:39 crc kubenswrapper[4706]: I1208 19:22:39.968796 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:39Z","lastTransitionTime":"2025-12-08T19:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.071750 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.071806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.071818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.071840 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.071860 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.174949 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.175007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.175017 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.175035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.175046 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.279126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.279175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.279184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.279203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.279220 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.381290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.381340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.381352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.381369 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.381381 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.483434 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.483472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.483481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.483496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.483505 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.586675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.586722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.586734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.586748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.586760 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.607236 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:40 crc kubenswrapper[4706]: E1208 19:22:40.607388 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.607466 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.607512 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:40 crc kubenswrapper[4706]: E1208 19:22:40.607699 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:40 crc kubenswrapper[4706]: E1208 19:22:40.607941 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.692933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.693004 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.693024 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.693053 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.693079 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.796684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.797145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.797250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.797364 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.797449 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.900352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.900711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.900834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.900924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:40 crc kubenswrapper[4706]: I1208 19:22:40.901003 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:40Z","lastTransitionTime":"2025-12-08T19:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.004081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.004114 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.004125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.004140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.004149 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.106550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.106886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.106955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.107054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.107150 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.210238 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.210305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.210317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.210336 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.210350 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.313379 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.314355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.314542 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.314686 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.314817 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.418213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.418311 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.418331 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.418359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.418377 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.521219 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.521350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.521380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.521422 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.521448 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.608198 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:41 crc kubenswrapper[4706]: E1208 19:22:41.608483 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.624119 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.624620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.624831 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.624966 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.625084 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.729175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.729582 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.729741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.729848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.729914 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.834510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.834924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.835108 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.835326 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.835522 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.938787 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.939243 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.939597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.939865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:41 crc kubenswrapper[4706]: I1208 19:22:41.940058 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:41Z","lastTransitionTime":"2025-12-08T19:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.043843 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.044373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.044768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.044961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.045117 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.148130 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.148663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.148907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.149071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.149214 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.253282 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.253347 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.253361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.253385 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.253411 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.356655 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.356720 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.356740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.356766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.356782 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.459676 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.459723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.459732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.459749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.459767 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.563778 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.563842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.563860 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.563887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.563905 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.607369 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.607404 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.607417 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:42 crc kubenswrapper[4706]: E1208 19:22:42.607561 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:42 crc kubenswrapper[4706]: E1208 19:22:42.607804 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:42 crc kubenswrapper[4706]: E1208 19:22:42.607841 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.666478 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.666543 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.666561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.666658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.666681 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.770417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.770500 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.770519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.770547 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.770572 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.873859 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.873933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.873998 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.874033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.874050 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.976812 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.976869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.976881 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.976902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:42 crc kubenswrapper[4706]: I1208 19:22:42.976917 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:42Z","lastTransitionTime":"2025-12-08T19:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.079621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.079680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.079695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.079722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.079736 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.183482 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.183541 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.183559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.183581 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.183594 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.287348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.287419 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.287428 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.287448 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.287460 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.342430 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:43 crc kubenswrapper[4706]: E1208 19:22:43.342639 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:43 crc kubenswrapper[4706]: E1208 19:22:43.342978 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs podName:439dd590-d737-435b-80a7-01f75fe6b64d nodeName:}" failed. No retries permitted until 2025-12-08 19:22:51.342955172 +0000 UTC m=+53.985156175 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs") pod "network-metrics-daemon-strs6" (UID: "439dd590-d737-435b-80a7-01f75fe6b64d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.390049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.390113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.390133 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.390157 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.390177 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.493747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.493829 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.493914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.493945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.493970 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.596164 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.596202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.596214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.596232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.596243 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.607848 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:43 crc kubenswrapper[4706]: E1208 19:22:43.608039 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.699325 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.699387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.699404 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.699426 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.699445 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.802789 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.802846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.802858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.802874 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.802886 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.905971 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.906034 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.906063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.906103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:43 crc kubenswrapper[4706]: I1208 19:22:43.906121 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:43Z","lastTransitionTime":"2025-12-08T19:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.010181 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.010287 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.010307 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.010335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.010358 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.112684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.112735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.112749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.112769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.112782 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.215498 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.215840 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.215977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.216084 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.216203 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.319744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.319830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.319856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.319886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.319907 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.422815 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.422882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.422899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.422961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.422987 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.525921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.525966 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.525977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.525993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.526004 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.607380 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.607490 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:44 crc kubenswrapper[4706]: E1208 19:22:44.607588 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.607490 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:44 crc kubenswrapper[4706]: E1208 19:22:44.607759 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:44 crc kubenswrapper[4706]: E1208 19:22:44.607876 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.629609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.629690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.629705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.629726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.629760 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.733055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.733129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.733149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.733178 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.733203 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.836591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.836668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.836687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.836717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.836739 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.940947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.941041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.941065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.941150 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:44 crc kubenswrapper[4706]: I1208 19:22:44.941178 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:44Z","lastTransitionTime":"2025-12-08T19:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.044981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.045041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.045050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.045065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.045075 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.150361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.150459 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.150493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.150535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.150562 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.253879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.253937 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.253951 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.253969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.253980 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.357853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.357984 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.358012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.358054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.358080 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.461120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.461197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.461215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.461253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.461304 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.573077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.573157 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.573175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.573200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.573219 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.607727 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:45 crc kubenswrapper[4706]: E1208 19:22:45.608044 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.676838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.676901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.676920 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.676947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.676967 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.780311 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.780771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.780950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.781097 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.781305 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.883563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.883975 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.884094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.884229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.884396 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.987741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.988111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.988244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.988418 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:45 crc kubenswrapper[4706]: I1208 19:22:45.988551 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:45Z","lastTransitionTime":"2025-12-08T19:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.091844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.091916 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.091934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.091961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.091983 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.195131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.195168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.195177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.195190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.195201 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.301192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.301247 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.301282 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.301302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.301315 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.404304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.404364 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.404374 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.404394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.404407 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.507405 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.507461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.507476 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.507496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.507508 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.607476 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.607602 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.607476 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:46 crc kubenswrapper[4706]: E1208 19:22:46.607713 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:46 crc kubenswrapper[4706]: E1208 19:22:46.607839 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:46 crc kubenswrapper[4706]: E1208 19:22:46.607994 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.613450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.613502 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.613519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.613542 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.613564 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.716695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.716742 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.716754 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.716771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.716785 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.821705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.821786 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.821803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.821830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.821848 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.925426 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.925493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.925510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.925541 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:46 crc kubenswrapper[4706]: I1208 19:22:46.925562 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:46Z","lastTransitionTime":"2025-12-08T19:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.028915 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.029006 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.029029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.029058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.029079 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.132810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.132894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.132918 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.132951 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.132977 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.236647 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.236698 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.236710 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.236730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.236743 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.339686 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.339757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.339783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.339815 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.339841 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.442385 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.442444 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.442455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.442472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.442483 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.545353 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.545415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.545433 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.545464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.545486 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.608623 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:47 crc kubenswrapper[4706]: E1208 19:22:47.608857 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.635509 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.648363 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.648440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.648459 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.648485 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.648503 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.660735 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.683521 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.708670 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.748677 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.752099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.752149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.752169 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.752197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.752217 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.779305 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.797685 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.822634 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.840702 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.854697 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.854738 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.854751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.854770 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.854783 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.864606 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.883173 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.905811 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.921681 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.938327 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.955843 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.957689 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.957739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.957750 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.957770 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.957781 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:47Z","lastTransitionTime":"2025-12-08T19:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.977382 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:47 crc kubenswrapper[4706]: I1208 19:22:47.992152 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:47Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.060588 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.060655 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.060668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.060689 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.060705 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.164197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.164318 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.164334 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.164356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.164369 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.267884 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.267953 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.267974 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.268001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.268018 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.371024 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.371109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.371128 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.371160 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.371205 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.473937 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.474344 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.474465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.474558 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.474626 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.506676 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.506879 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.506926 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.506964 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.507022 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507207 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507288 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507347 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:23:20.507288216 +0000 UTC m=+83.149489219 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507386 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:23:20.507375608 +0000 UTC m=+83.149576611 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507397 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507429 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:23:20.507401779 +0000 UTC m=+83.149602982 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507429 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507468 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507509 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 19:23:20.507500631 +0000 UTC m=+83.149701854 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507740 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507813 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.507895 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.508026 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 19:23:20.508001795 +0000 UTC m=+83.150202798 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.578602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.578655 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.578671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.578696 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.578714 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.607775 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.607838 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.608288 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.607920 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.608380 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:48 crc kubenswrapper[4706]: E1208 19:22:48.608487 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.681980 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.682059 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.682071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.682116 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.682133 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.784887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.784925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.784935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.784950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.784962 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.887509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.887847 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.887907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.887976 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.888035 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.990905 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.990970 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.990992 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.991020 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:48 crc kubenswrapper[4706]: I1208 19:22:48.991040 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:48Z","lastTransitionTime":"2025-12-08T19:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.094664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.094760 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.094776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.094822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.094837 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.197776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.197834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.197846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.197866 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.197879 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.301044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.301465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.301568 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.301664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.301755 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.405106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.405158 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.405169 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.405187 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.405199 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.507628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.507665 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.507675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.507689 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.507700 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.532888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.533346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.533563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.533755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.533989 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: E1208 19:22:49.556520 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.561827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.561867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.561875 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.561891 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.561903 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: E1208 19:22:49.577715 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.584127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.584255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.584318 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.584356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.584381 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: E1208 19:22:49.605187 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.608464 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:49 crc kubenswrapper[4706]: E1208 19:22:49.608664 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.611172 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.611380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.611415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.611458 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.611486 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: E1208 19:22:49.637725 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.642736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.642790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.642802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.642821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.642851 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: E1208 19:22:49.661982 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: E1208 19:22:49.662119 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.664227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.664305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.664321 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.664338 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.664397 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.767480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.767526 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.767539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.767560 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.767574 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.801166 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.817408 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.817482 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.831191 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.848169 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.867154 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.870532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.870598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.870613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.870630 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.870642 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.882945 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.899163 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.921473 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.937995 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.960528 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.973662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.973756 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.973780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.973816 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.973840 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:49Z","lastTransitionTime":"2025-12-08T19:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.979255 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:49 crc kubenswrapper[4706]: I1208 19:22:49.997698 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:49Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.017218 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:50Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.040169 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:50Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.073930 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:50Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.077927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.078167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.078329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.078484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.078613 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.090030 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:50Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.109989 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:50Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.123803 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:50Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.182974 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.183051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.183070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.183099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.183123 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.287351 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.287399 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.287412 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.287433 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.287445 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.390613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.390723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.390740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.390761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.390774 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.494842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.494895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.494910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.494927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.494940 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.597579 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.597633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.597646 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.597668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.597686 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.608013 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.608157 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.608025 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:50 crc kubenswrapper[4706]: E1208 19:22:50.608304 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:50 crc kubenswrapper[4706]: E1208 19:22:50.608466 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:50 crc kubenswrapper[4706]: E1208 19:22:50.608626 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.700966 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.701023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.701035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.701056 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.701069 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.804381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.804715 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.804822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.804950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.805041 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.908433 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.908781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.908871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.908976 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:50 crc kubenswrapper[4706]: I1208 19:22:50.909068 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:50Z","lastTransitionTime":"2025-12-08T19:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.012298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.012363 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.012378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.012399 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.012419 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.115307 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.115362 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.115380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.115403 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.115420 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.218389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.218443 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.218456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.218480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.218495 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.321337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.321741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.321762 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.321791 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.321812 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.425145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.425200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.425213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.425232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.425246 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.442983 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:51 crc kubenswrapper[4706]: E1208 19:22:51.443235 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:51 crc kubenswrapper[4706]: E1208 19:22:51.443376 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs podName:439dd590-d737-435b-80a7-01f75fe6b64d nodeName:}" failed. No retries permitted until 2025-12-08 19:23:07.443340868 +0000 UTC m=+70.085542041 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs") pod "network-metrics-daemon-strs6" (UID: "439dd590-d737-435b-80a7-01f75fe6b64d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.529144 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.529210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.529228 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.529254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.529304 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.608201 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:51 crc kubenswrapper[4706]: E1208 19:22:51.608830 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.609309 4706 scope.go:117] "RemoveContainer" containerID="1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.634570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.634885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.635159 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.635427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.635667 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.738963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.739005 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.739013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.739029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.739040 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.842674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.842744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.842761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.842806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.842826 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.946232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.946795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.946806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.946926 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:51 crc kubenswrapper[4706]: I1208 19:22:51.946941 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:51Z","lastTransitionTime":"2025-12-08T19:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.049430 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.049471 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.049481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.049502 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.049515 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.088305 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/1.log" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.091061 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.091572 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.104580 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.115421 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.141471 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.153418 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.153461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.153471 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.153489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.153503 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.160898 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.176385 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.190702 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.203760 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.215804 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.228080 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.243087 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.256319 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.256644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.256797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.256892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.256987 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.257924 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.271569 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.285684 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.300558 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.316098 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.343461 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.359637 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.359831 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.359866 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.359881 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.359900 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.359914 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.376239 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:52Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.464112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.464759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.464779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.464808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.464829 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.576586 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.576659 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.576685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.576721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.576748 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.607814 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:52 crc kubenswrapper[4706]: E1208 19:22:52.608119 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.607934 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:52 crc kubenswrapper[4706]: E1208 19:22:52.608889 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.607853 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:52 crc kubenswrapper[4706]: E1208 19:22:52.609236 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.679981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.680046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.680058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.680102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.680117 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.783401 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.783465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.783488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.783517 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.783536 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.890687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.890733 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.890769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.890788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.890803 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.994985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.995031 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.995045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.995064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:52 crc kubenswrapper[4706]: I1208 19:22:52.995077 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:52Z","lastTransitionTime":"2025-12-08T19:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.097765 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/2.log" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.098158 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.098217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.098235 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.098293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.098313 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.098870 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/1.log" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.102591 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614" exitCode=1 Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.102657 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.102720 4706 scope.go:117] "RemoveContainer" containerID="1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.103583 4706 scope.go:117] "RemoveContainer" containerID="733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614" Dec 08 19:22:53 crc kubenswrapper[4706]: E1208 19:22:53.103823 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.120634 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.135168 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.162704 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.178053 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.193815 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.200810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.200854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.200868 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.200892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.200908 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.206445 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.220239 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.234168 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.247340 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.265436 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.277048 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.289722 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.302837 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.304294 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.304324 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.304335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.304381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.304394 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.317814 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.336996 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.356968 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a5a937b24d12398425554cec3d3208bd41fae8981d4988731fac07e16fceb91\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:33Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:32Z is after 2025-08-24T17:21:41Z]\\\\nI1208 19:22:32.911073 6143 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00777ecbf \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.371582 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.387981 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:53Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.407002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.407052 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.407065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.407086 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.407099 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.509561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.509612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.509622 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.509638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.509647 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.608324 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:53 crc kubenswrapper[4706]: E1208 19:22:53.608518 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.612011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.612035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.612044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.612060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.612072 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.714663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.714703 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.714712 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.714726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.714737 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.817740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.817785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.817794 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.817812 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.817822 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.921245 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.921313 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.921324 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.921340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:53 crc kubenswrapper[4706]: I1208 19:22:53.921354 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:53Z","lastTransitionTime":"2025-12-08T19:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.024182 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.024227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.024237 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.024257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.024285 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.109929 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/2.log" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.114557 4706 scope.go:117] "RemoveContainer" containerID="733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614" Dec 08 19:22:54 crc kubenswrapper[4706]: E1208 19:22:54.114732 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.126811 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.126861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.126872 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.126888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.126902 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.149033 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.169930 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.190240 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.206554 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.223575 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.229514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.229574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.229603 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.229637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.229663 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.237844 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.255125 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.270693 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.284341 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.304404 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.315163 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.327569 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.331900 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.332176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.332283 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.332411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.332496 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.339758 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.357517 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.370820 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.383485 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.400129 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.415555 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:54Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.435671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.435706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.435716 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.435732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.435742 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.539196 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.539744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.539891 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.540050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.540197 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.608093 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.608093 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.608141 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:54 crc kubenswrapper[4706]: E1208 19:22:54.608843 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:54 crc kubenswrapper[4706]: E1208 19:22:54.608984 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:54 crc kubenswrapper[4706]: E1208 19:22:54.608666 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.644361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.644496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.644522 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.644561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.644588 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.748224 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.748319 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.748340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.748365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.748384 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.851613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.852065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.852241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.852501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.852658 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.956175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.956754 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.957008 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.957234 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:54 crc kubenswrapper[4706]: I1208 19:22:54.957496 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:54Z","lastTransitionTime":"2025-12-08T19:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.060631 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.061015 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.061159 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.061391 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.061604 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.165806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.165899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.165925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.165957 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.165978 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.270465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.270562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.270588 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.270627 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.270657 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.373872 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.373933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.373947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.373967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.373983 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.477549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.477714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.477735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.477764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.477790 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.580973 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.581056 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.581080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.581113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.581139 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.607726 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:55 crc kubenswrapper[4706]: E1208 19:22:55.607898 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.684386 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.684448 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.684465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.684492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.684511 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.787982 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.788067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.788093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.788123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.788144 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.891427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.891464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.891473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.891488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.891500 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.994767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.994815 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.994824 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.994839 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:55 crc kubenswrapper[4706]: I1208 19:22:55.994848 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:55Z","lastTransitionTime":"2025-12-08T19:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.097923 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.098015 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.098036 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.098067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.098087 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.202611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.202687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.202712 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.202744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.202771 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.305633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.305679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.305689 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.305706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.305719 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.408611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.408671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.408688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.408716 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.408734 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.517525 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.517814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.517922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.517994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.518054 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.607920 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.607980 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.607918 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:56 crc kubenswrapper[4706]: E1208 19:22:56.608190 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:56 crc kubenswrapper[4706]: E1208 19:22:56.608382 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:56 crc kubenswrapper[4706]: E1208 19:22:56.608497 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.620985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.621055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.621077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.621111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.621134 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.724763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.724828 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.724852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.724885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.724910 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.828859 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.828936 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.828964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.828997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.829025 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.932633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.933085 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.933290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.933450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:56 crc kubenswrapper[4706]: I1208 19:22:56.933588 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:56Z","lastTransitionTime":"2025-12-08T19:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.037142 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.037569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.037650 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.037757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.037839 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.140981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.141075 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.141096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.141126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.141144 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.244503 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.244594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.244618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.244657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.244682 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.348548 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.348641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.348661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.348691 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.348709 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.451623 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.451679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.451694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.451715 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.451730 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.553936 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.554001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.554015 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.554037 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.554054 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.607343 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:57 crc kubenswrapper[4706]: E1208 19:22:57.607763 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.620992 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.633230 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.645413 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.656967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.657016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.657027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.657044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.657056 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.669749 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.682677 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.694049 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.708850 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.722348 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.735017 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.750660 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.759372 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.759413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.759424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.759441 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.759451 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.766402 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.780826 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.799224 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.815725 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.836393 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.850794 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.862671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.863179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.863506 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.863741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.863959 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.874289 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.890466 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:57Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.966678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.966711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.966719 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.966736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:57 crc kubenswrapper[4706]: I1208 19:22:57.966747 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:57Z","lastTransitionTime":"2025-12-08T19:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.068735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.068773 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.068783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.068797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.068807 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.172009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.172066 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.172083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.172108 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.172127 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.275547 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.275592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.275601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.275621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.275633 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.379096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.379143 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.379151 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.379170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.379184 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.483427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.483485 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.483502 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.483524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.483539 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.587322 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.587375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.587389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.587411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.587434 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.607630 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.607683 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.607811 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:22:58 crc kubenswrapper[4706]: E1208 19:22:58.607815 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:22:58 crc kubenswrapper[4706]: E1208 19:22:58.607938 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:22:58 crc kubenswrapper[4706]: E1208 19:22:58.608036 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.694987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.695053 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.695068 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.695090 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.695108 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.798645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.798717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.798730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.798751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.798764 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.901523 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.901577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.901606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.901628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:58 crc kubenswrapper[4706]: I1208 19:22:58.901642 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:58Z","lastTransitionTime":"2025-12-08T19:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.004446 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.004560 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.004574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.004598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.004614 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.108256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.108492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.108528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.108564 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.108583 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.211374 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.211446 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.211465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.211495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.211518 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.314550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.314621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.314639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.314668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.314685 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.418328 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.418400 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.418419 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.418448 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.418467 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.521381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.521422 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.521434 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.521450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.521460 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.607934 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:22:59 crc kubenswrapper[4706]: E1208 19:22:59.608139 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.624481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.624546 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.624562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.624594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.624612 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.727969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.728092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.728114 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.728143 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.728214 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.827315 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.827385 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.827408 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.827446 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.827467 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: E1208 19:22:59.847153 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:59Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.852595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.852645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.852661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.852683 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.852698 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: E1208 19:22:59.868314 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:59Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.873150 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.873240 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.873301 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.873343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.873388 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: E1208 19:22:59.890352 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:59Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.895999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.896037 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.896049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.896071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.896082 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: E1208 19:22:59.913930 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:59Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.919170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.919332 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.919356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.919379 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.919394 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:22:59 crc kubenswrapper[4706]: E1208 19:22:59.935354 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:22:59Z is after 2025-08-24T17:21:41Z" Dec 08 19:22:59 crc kubenswrapper[4706]: E1208 19:22:59.935586 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.937811 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.937844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.937858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.937877 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:22:59 crc kubenswrapper[4706]: I1208 19:22:59.937890 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:22:59Z","lastTransitionTime":"2025-12-08T19:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.040303 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.040355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.040371 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.040392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.040408 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.143377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.143456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.143466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.143484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.143501 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.245985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.246034 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.246044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.246064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.246074 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.349575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.350013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.350034 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.350057 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.350072 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.452917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.452975 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.452989 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.453012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.453028 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.556503 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.556554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.556564 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.556595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.556611 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.608077 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.608156 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.608240 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:00 crc kubenswrapper[4706]: E1208 19:23:00.608311 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:00 crc kubenswrapper[4706]: E1208 19:23:00.608422 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:00 crc kubenswrapper[4706]: E1208 19:23:00.608531 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.664775 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.665617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.665650 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.665696 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.665726 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.769062 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.769102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.769110 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.769125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.769136 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.871553 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.871611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.871624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.871643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.871657 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.975811 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.975867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.975879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.975901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:00 crc kubenswrapper[4706]: I1208 19:23:00.975913 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:00Z","lastTransitionTime":"2025-12-08T19:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.079167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.079684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.079853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.080002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.080145 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.183671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.183725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.183735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.183753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.183764 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.286349 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.286387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.286395 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.286408 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.286417 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.389046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.389099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.389111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.389132 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.389154 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.491999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.492051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.492060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.492077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.492088 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.595038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.595096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.595106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.595120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.595132 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.607367 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:01 crc kubenswrapper[4706]: E1208 19:23:01.607552 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.697912 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.697959 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.697968 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.698057 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.698067 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.800864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.800921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.800934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.800950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.800961 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.904440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.904497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.904509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.904525 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:01 crc kubenswrapper[4706]: I1208 19:23:01.904538 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:01Z","lastTransitionTime":"2025-12-08T19:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.008014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.008078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.008094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.008121 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.008137 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.111332 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.111382 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.111394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.111413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.111425 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.214337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.214370 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.214380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.214395 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.214405 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.317416 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.317462 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.317476 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.317492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.317505 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.420917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.420977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.420990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.421023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.421042 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.524554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.524605 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.524617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.524633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.524648 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.607561 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.607595 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.607595 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:02 crc kubenswrapper[4706]: E1208 19:23:02.607706 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:02 crc kubenswrapper[4706]: E1208 19:23:02.607840 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:02 crc kubenswrapper[4706]: E1208 19:23:02.607914 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.626852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.626922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.626933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.626955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.626969 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.729256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.729342 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.729352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.729370 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.729406 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.832451 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.832498 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.832509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.832523 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.832532 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.935032 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.935078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.935089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.935103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:02 crc kubenswrapper[4706]: I1208 19:23:02.935113 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:02Z","lastTransitionTime":"2025-12-08T19:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.037790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.037853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.037868 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.037888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.037903 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.141161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.141216 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.141229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.141250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.141280 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.243928 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.243968 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.243983 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.243999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.244011 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.346853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.346945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.346957 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.346974 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.347003 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.449629 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.449668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.449678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.449695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.449707 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.552535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.552591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.552600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.552621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.552647 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.607357 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:03 crc kubenswrapper[4706]: E1208 19:23:03.607530 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.655500 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.655553 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.655565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.655581 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.655594 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.758374 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.758448 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.758461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.758477 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.758492 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.861499 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.861578 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.861595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.861621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.861644 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.964054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.964091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.964102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.964118 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:03 crc kubenswrapper[4706]: I1208 19:23:03.964128 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:03Z","lastTransitionTime":"2025-12-08T19:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.067250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.067330 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.067339 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.067378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.067392 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.170085 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.170143 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.170157 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.170179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.170191 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.272673 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.272731 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.272746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.272767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.272782 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.375121 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.375162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.375171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.375187 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.375197 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.477763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.477810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.477821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.477838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.477854 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.580039 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.580073 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.580082 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.580134 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.580146 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.607654 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.607677 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.607712 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:04 crc kubenswrapper[4706]: E1208 19:23:04.607761 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:04 crc kubenswrapper[4706]: E1208 19:23:04.607843 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:04 crc kubenswrapper[4706]: E1208 19:23:04.607910 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.682838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.682895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.682905 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.682924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.682937 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.786820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.786867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.786879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.786898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.786911 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.889684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.889722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.889731 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.889744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.889754 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.992699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.992756 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.992768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.992786 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:04 crc kubenswrapper[4706]: I1208 19:23:04.992799 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:04Z","lastTransitionTime":"2025-12-08T19:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.096644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.096699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.096713 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.096736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.096750 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.199444 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.199503 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.199525 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.199546 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.199559 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.302440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.302514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.302532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.302561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.302592 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.405465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.405925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.406049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.406168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.406250 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.508853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.508898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.508907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.508927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.508937 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.607328 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:05 crc kubenswrapper[4706]: E1208 19:23:05.607501 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.608280 4706 scope.go:117] "RemoveContainer" containerID="733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614" Dec 08 19:23:05 crc kubenswrapper[4706]: E1208 19:23:05.608442 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.611955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.612016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.612035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.612058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.612077 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.714781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.714827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.714839 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.714856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.714868 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.818070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.818159 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.818173 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.818252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.818296 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.921142 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.921192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.921204 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.921228 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:05 crc kubenswrapper[4706]: I1208 19:23:05.921247 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:05Z","lastTransitionTime":"2025-12-08T19:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.024663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.024743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.024764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.024796 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.024819 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.128058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.128125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.128140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.128163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.128177 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.231812 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.231881 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.231894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.231919 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.231933 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.334198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.334247 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.334277 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.334298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.334311 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.436714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.436770 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.436783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.436805 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.436816 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.539882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.539961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.539980 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.540009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.540044 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.607981 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.608036 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.607988 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:06 crc kubenswrapper[4706]: E1208 19:23:06.608202 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:06 crc kubenswrapper[4706]: E1208 19:23:06.608354 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:06 crc kubenswrapper[4706]: E1208 19:23:06.608489 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.643165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.643207 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.643217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.643233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.643244 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.745980 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.746047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.746058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.746078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.746095 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.848787 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.848839 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.848854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.848878 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.848893 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.951302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.951347 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.951359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.951376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:06 crc kubenswrapper[4706]: I1208 19:23:06.951387 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:06Z","lastTransitionTime":"2025-12-08T19:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.054345 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.055021 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.055119 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.055195 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.055274 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.159192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.159669 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.159854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.160045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.160315 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.263478 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.263531 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.263544 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.263564 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.263579 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.366609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.367038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.367212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.367342 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.367444 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.444767 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:07 crc kubenswrapper[4706]: E1208 19:23:07.445095 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:23:07 crc kubenswrapper[4706]: E1208 19:23:07.445249 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs podName:439dd590-d737-435b-80a7-01f75fe6b64d nodeName:}" failed. No retries permitted until 2025-12-08 19:23:39.445212222 +0000 UTC m=+102.087413375 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs") pod "network-metrics-daemon-strs6" (UID: "439dd590-d737-435b-80a7-01f75fe6b64d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.470251 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.470314 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.470326 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.470345 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.470359 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.573176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.573234 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.573251 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.573316 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.573337 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.607784 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:07 crc kubenswrapper[4706]: E1208 19:23:07.607963 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.630729 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.648994 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.664514 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.676742 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.676796 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.676810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.676833 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.676847 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.680237 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.695377 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.715142 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.730803 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.743750 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.753200 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.763291 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.774316 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.778917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.779109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.779176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.779272 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.779368 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.788848 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.807767 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.825851 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.841960 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.854213 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.868377 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.879289 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:07Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.882418 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.882493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.882531 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.882555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.882569 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.986432 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.986495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.986508 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.986530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:07 crc kubenswrapper[4706]: I1208 19:23:07.986544 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:07Z","lastTransitionTime":"2025-12-08T19:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.089699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.089766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.089788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.089819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.089840 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.170932 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/0.log" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.170999 4706 generic.go:334] "Generic (PLEG): container finished" podID="31d41aa9-3a74-42f4-9933-3f484a89c41d" containerID="92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5" exitCode=1 Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.171042 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x674x" event={"ID":"31d41aa9-3a74-42f4-9933-3f484a89c41d","Type":"ContainerDied","Data":"92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.171543 4706 scope.go:117] "RemoveContainer" containerID="92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.187431 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.192386 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.192414 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.192425 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.192442 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.192454 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.202738 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.217363 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.232824 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.248078 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.266687 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.279702 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.295717 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.295922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.295965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.295978 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.296000 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.296013 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.311286 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.332844 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.348291 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.366043 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.382736 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.398176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.398227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.398241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.398274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.398288 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.398466 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:23:07Z\\\",\\\"message\\\":\\\"2025-12-08T19:22:22+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3\\\\n2025-12-08T19:22:22+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3 to /host/opt/cni/bin/\\\\n2025-12-08T19:22:22Z [verbose] multus-daemon started\\\\n2025-12-08T19:22:22Z [verbose] Readiness Indicator file check\\\\n2025-12-08T19:23:07Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.418071 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.432594 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.447854 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.460515 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:08Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.500907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.500958 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.500967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.500986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.500996 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.603933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.603987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.604000 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.604022 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.604037 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.608280 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.608300 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.608282 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:08 crc kubenswrapper[4706]: E1208 19:23:08.608422 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:08 crc kubenswrapper[4706]: E1208 19:23:08.608480 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:08 crc kubenswrapper[4706]: E1208 19:23:08.608557 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.705948 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.705986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.706000 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.706019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.706030 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.807990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.808047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.808064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.808088 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.808105 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.910956 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.911016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.911028 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.911050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:08 crc kubenswrapper[4706]: I1208 19:23:08.911062 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:08Z","lastTransitionTime":"2025-12-08T19:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.014380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.014445 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.014458 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.014484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.014502 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.117355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.117435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.117449 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.117466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.117479 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.175958 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/0.log" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.176030 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x674x" event={"ID":"31d41aa9-3a74-42f4-9933-3f484a89c41d","Type":"ContainerStarted","Data":"f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.191991 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.207280 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.220680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.220726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.220736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.220756 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.220767 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.220924 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.236813 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.251290 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.262290 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.275832 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.294478 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.312201 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.323081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.323142 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.323158 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.323177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.323193 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.327660 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.342545 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.358152 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:23:07Z\\\",\\\"message\\\":\\\"2025-12-08T19:22:22+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3\\\\n2025-12-08T19:22:22+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3 to /host/opt/cni/bin/\\\\n2025-12-08T19:22:22Z [verbose] multus-daemon started\\\\n2025-12-08T19:22:22Z [verbose] Readiness Indicator file check\\\\n2025-12-08T19:23:07Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:23:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.379092 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.398245 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.408329 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.419771 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.425726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.425771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.425789 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.425811 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.425829 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.431094 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.441454 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:09Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.528363 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.528421 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.528435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.528459 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.528476 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.608476 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:09 crc kubenswrapper[4706]: E1208 19:23:09.608704 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.631534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.631585 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.631594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.631607 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.631617 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.735474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.735536 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.735556 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.735577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.735591 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.838831 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.838894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.838915 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.838942 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.838960 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.941938 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.941986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.941996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.942014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:09 crc kubenswrapper[4706]: I1208 19:23:09.942025 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:09Z","lastTransitionTime":"2025-12-08T19:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.044822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.044869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.044877 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.044896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.044908 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.048608 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.048640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.048648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.048659 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.048668 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.064001 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:10Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.071873 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.071922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.071932 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.071961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.071973 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.087527 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:10Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.091048 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.091070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.091078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.091091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.091101 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.103394 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:10Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.107257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.107318 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.107328 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.107343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.107354 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.118311 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:10Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.125757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.125802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.125810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.125826 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.125839 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.137927 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cf23845b-400b-4cbc-a718-15bd67773250\\\",\\\"systemUUID\\\":\\\"fcc1c5cb-e33f-45eb-b8f8-156e819c7d80\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:10Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.138119 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.147703 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.147753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.147766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.147786 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.147799 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.250163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.250200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.250210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.250722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.250749 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.353795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.354197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.354334 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.354417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.354480 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.457692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.457738 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.457749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.457767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.457778 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.560376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.560728 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.560862 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.560989 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.561079 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.607907 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.608381 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.608124 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.608065 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.608665 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:10 crc kubenswrapper[4706]: E1208 19:23:10.608881 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.620455 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.664550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.664593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.664602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.664617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.664627 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.767684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.767996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.768100 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.768199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.768311 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.871636 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.871680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.871688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.871705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.871718 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.974715 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.974765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.974780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.974801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:10 crc kubenswrapper[4706]: I1208 19:23:10.974815 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:10Z","lastTransitionTime":"2025-12-08T19:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.077925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.077965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.077976 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.077993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.078005 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.181329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.182154 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.182246 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.182364 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.182446 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.285801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.285846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.285858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.285872 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.285884 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.388602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.388672 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.388697 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.388726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.388750 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.492196 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.492249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.492297 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.492331 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.492346 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.596642 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.596731 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.596771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.596809 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.596836 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.608328 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:11 crc kubenswrapper[4706]: E1208 19:23:11.608500 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.699435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.699474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.699484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.699499 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.699509 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.802308 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.802339 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.802346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.802360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.802397 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.905215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.905278 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.905295 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.905315 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:11 crc kubenswrapper[4706]: I1208 19:23:11.905331 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:11Z","lastTransitionTime":"2025-12-08T19:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.014489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.014588 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.014615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.014648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.014672 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.117865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.118366 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.118692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.118891 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.119043 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.223070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.224221 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.224414 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.224559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.224682 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.328892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.329450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.329717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.329921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.330086 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.433996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.434608 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.434806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.435009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.435156 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.538820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.538886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.538905 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.538963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.538984 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.608174 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:12 crc kubenswrapper[4706]: E1208 19:23:12.608796 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.608402 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.608303 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:12 crc kubenswrapper[4706]: E1208 19:23:12.609317 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:12 crc kubenswrapper[4706]: E1208 19:23:12.609562 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.643460 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.643545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.643565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.643593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.643614 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.747682 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.747771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.747783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.747802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.747815 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.850555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.850612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.850623 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.850640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.850652 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.953659 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.953714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.953724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.953745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:12 crc kubenswrapper[4706]: I1208 19:23:12.953757 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:12Z","lastTransitionTime":"2025-12-08T19:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.057096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.057179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.057199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.057227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.057248 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.161101 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.161191 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.161217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.161255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.161317 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.265094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.265180 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.265206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.265237 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.265309 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.369578 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.369650 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.369673 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.369710 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.369735 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.473148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.473221 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.473245 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.473330 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.473356 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.576522 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.576576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.576590 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.576611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.576621 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.608250 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:13 crc kubenswrapper[4706]: E1208 19:23:13.608432 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.680425 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.680494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.680512 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.680539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.680557 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.784238 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.784319 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.784335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.784360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.784377 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.888064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.888141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.888174 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.888197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.888213 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.990950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.991047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.991116 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.991150 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:13 crc kubenswrapper[4706]: I1208 19:23:13.991208 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:13Z","lastTransitionTime":"2025-12-08T19:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.094760 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.094828 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.094841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.094867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.094886 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.198304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.198599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.198738 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.198837 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.198925 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.303253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.304492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.304691 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.304878 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.305061 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.409071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.409578 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.409788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.410047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.410292 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.514602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.515127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.515380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.515600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.515796 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.608608 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.608644 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:14 crc kubenswrapper[4706]: E1208 19:23:14.608938 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.608644 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:14 crc kubenswrapper[4706]: E1208 19:23:14.609066 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:14 crc kubenswrapper[4706]: E1208 19:23:14.609252 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.619478 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.619546 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.619567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.619597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.619618 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.723969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.724042 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.724063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.724100 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.724120 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.826934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.826994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.827007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.827027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.827041 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.930413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.930482 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.930511 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.930545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:14 crc kubenswrapper[4706]: I1208 19:23:14.930568 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:14Z","lastTransitionTime":"2025-12-08T19:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.035396 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.035476 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.035495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.035539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.035561 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.138150 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.138213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.138228 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.138248 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.138283 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.241035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.241089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.241109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.241131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.241147 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.344213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.344293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.344306 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.344332 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.344348 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.447705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.447772 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.447792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.447818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.447837 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.551194 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.551252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.551318 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.551338 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.551350 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.608321 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:15 crc kubenswrapper[4706]: E1208 19:23:15.608535 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.654680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.654755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.654783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.654820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.654850 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.759249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.759362 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.759381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.759408 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.759428 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.862246 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.862375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.862407 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.862445 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.862476 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.965069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.965127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.965148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.965174 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:15 crc kubenswrapper[4706]: I1208 19:23:15.965190 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:15Z","lastTransitionTime":"2025-12-08T19:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.068341 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.068436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.068461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.068497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.068521 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.172502 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.172571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.172580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.172598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.172610 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.276046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.276480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.276532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.276574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.276605 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.380162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.380214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.380223 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.380238 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.380253 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.483497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.483569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.483587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.483612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.483632 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.587381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.587434 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.587443 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.587466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.587477 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.607881 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.607994 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.608095 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:16 crc kubenswrapper[4706]: E1208 19:23:16.608143 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:16 crc kubenswrapper[4706]: E1208 19:23:16.608219 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:16 crc kubenswrapper[4706]: E1208 19:23:16.608423 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.609767 4706 scope.go:117] "RemoveContainer" containerID="733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.690972 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.691475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.691487 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.691504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.691517 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.795842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.795911 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.795928 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.795956 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.795976 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.898947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.899019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.899047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.899081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:16 crc kubenswrapper[4706]: I1208 19:23:16.899106 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:16Z","lastTransitionTime":"2025-12-08T19:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.002456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.002510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.002524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.002580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.002604 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.104991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.105040 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.105060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.105084 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.105099 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.206863 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.206928 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.206942 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.206964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.206982 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.209352 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/2.log" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.212206 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.212792 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.226510 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.239345 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.254910 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:23:07Z\\\",\\\"message\\\":\\\"2025-12-08T19:22:22+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3\\\\n2025-12-08T19:22:22+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3 to /host/opt/cni/bin/\\\\n2025-12-08T19:22:22Z [verbose] multus-daemon started\\\\n2025-12-08T19:22:22Z [verbose] Readiness Indicator file check\\\\n2025-12-08T19:23:07Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:23:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.274917 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:23:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.290127 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.303594 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.310001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.310047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.310062 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.310083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.310101 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.317605 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.333865 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.372683 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.386491 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.400501 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b3ef69-0fe7-4e09-9c2d-6c9c95e25540\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fbd3072ee050c9a83353949c6a2f3d761019bd054ae54a97ea97a9aea27fd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.412818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.412874 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.412888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.412910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.412925 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.414111 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.428458 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.440017 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.451926 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.469462 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.489368 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.503489 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.516391 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.516440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.516452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.516472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.516438 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.516483 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.607912 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:17 crc kubenswrapper[4706]: E1208 19:23:17.608131 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.618940 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.618994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.619007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.619027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.619041 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.628612 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.644678 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.667384 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.687773 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.705832 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b3ef69-0fe7-4e09-9c2d-6c9c95e25540\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fbd3072ee050c9a83353949c6a2f3d761019bd054ae54a97ea97a9aea27fd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.721594 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.721711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.721767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.721785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.721813 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.721834 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.741659 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.760167 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.776229 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.797288 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.813578 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.825155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.825221 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.825237 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.825258 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.825670 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.830948 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.847408 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.862222 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.873651 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.887163 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:23:07Z\\\",\\\"message\\\":\\\"2025-12-08T19:22:22+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3\\\\n2025-12-08T19:22:22+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3 to /host/opt/cni/bin/\\\\n2025-12-08T19:22:22Z [verbose] multus-daemon started\\\\n2025-12-08T19:22:22Z [verbose] Readiness Indicator file check\\\\n2025-12-08T19:23:07Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:23:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.910568 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:23:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.929491 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.929530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.929539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.929558 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.929568 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:17Z","lastTransitionTime":"2025-12-08T19:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.931639 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:17 crc kubenswrapper[4706]: I1208 19:23:17.954023 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:17Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.033700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.033776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.033801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.033834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.033859 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.136592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.136645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.136665 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.136690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.136707 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.219410 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/3.log" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.221665 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/2.log" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.227080 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" exitCode=1 Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.227137 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.227189 4706 scope.go:117] "RemoveContainer" containerID="733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.228748 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:23:18 crc kubenswrapper[4706]: E1208 19:23:18.229175 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.240153 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.240212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.240226 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.240248 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.240283 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.247991 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b3ef69-0fe7-4e09-9c2d-6c9c95e25540\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fbd3072ee050c9a83353949c6a2f3d761019bd054ae54a97ea97a9aea27fd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.265629 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.282618 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.299945 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.319581 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.344442 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.344498 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.344543 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.344562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.344575 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.348627 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.363591 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.383475 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.403488 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.425321 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.448302 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:23:07Z\\\",\\\"message\\\":\\\"2025-12-08T19:22:22+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3\\\\n2025-12-08T19:22:22+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3 to /host/opt/cni/bin/\\\\n2025-12-08T19:22:22Z [verbose] multus-daemon started\\\\n2025-12-08T19:22:22Z [verbose] Readiness Indicator file check\\\\n2025-12-08T19:23:07Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:23:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.454643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.454692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.454712 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.454743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.454765 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.487813 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733958e20bdeb6e4b32a9e71ea8ecfd27e402ccf911affa63babd9db16043614\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:22:52Z\\\",\\\"message\\\":\\\"k-metrics-daemon-strs6 openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-machine-config-operator/machine-config-daemon-kp47b openshift-network-node-identity/network-node-identity-vrzqb openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-qxrgm openshift-multus/multus-additional-cni-plugins-ll6w2 openshift-ovn-kubernetes/ovnkube-node-kl95h openshift-multus/multus-x674x openshift-network-operator/iptables-alerter-4ln5h]\\\\nF1208 19:22:52.413323 6384 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:23:17Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1208 19:23:17.629774 6753 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1208 19:23:17.629803 6753 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1208 19:23:17.629817 6753 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1208 19:23:17.629864 6753 factory.go:1336] Added *v1.Node event handler 7\\\\nI1208 19:23:17.629913 6753 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1208 19:23:17.630229 6753 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1208 19:23:17.630321 6753 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1208 19:23:17.630352 6753 ovnkube.go:599] Stopped ovnkube\\\\nI1208 19:23:17.630377 6753 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1208 19:23:17.630448 6753 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:23:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.514529 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.536074 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.556896 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.558234 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.558289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.558300 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.558319 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.558332 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.574424 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.608369 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.608436 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.608483 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:18 crc kubenswrapper[4706]: E1208 19:23:18.608590 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:18 crc kubenswrapper[4706]: E1208 19:23:18.608803 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:18 crc kubenswrapper[4706]: E1208 19:23:18.608990 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.613928 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.635105 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.657426 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:18Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.661667 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.661709 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.661725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.661749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.661770 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.765242 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.765323 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.765336 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.765354 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.765371 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.869817 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.869899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.869923 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.869955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.869977 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.973482 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.973555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.973567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.973606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:18 crc kubenswrapper[4706]: I1208 19:23:18.973624 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:18Z","lastTransitionTime":"2025-12-08T19:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.076374 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.076442 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.076462 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.076490 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.076511 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.180149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.180699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.180859 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.181013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.181157 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.234731 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/3.log" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.240029 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:23:19 crc kubenswrapper[4706]: E1208 19:23:19.240247 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.275091 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25608ebc-2194-4347-802f-963ee42aa369\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8ec5e1106139d3ddaec43d186af3f58f5b728c4a6c2ee763c9ba896904d353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a87cfd8cea690276f28b0f6dbd8664efc4e83003eb7c79f51568ed92866884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c5709c9ae24cd61ba0cb2c518e3fefa299d2381143a5962528cceeeff2bc99b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661852abd345672ca7400399e1f02cdc1592c72c75b9937083ac8b5b33565e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff5cd5e355f3aeaae6f66b9be21854b8237cff680aa732baa72d850dcf988e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc53ffa0ac98f84144471defe2640a08fad9ccd13baae71a4e880227cdf4f644\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fd0b5cd7c82390583ed6e5bf6cf6bd5d358c9dee1deb624e30b2eca03f4e9ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa9342ebb276f243d88186e6b57c9b7354f7a24d10ecd3ef9db1e9910f0787bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.284240 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.284303 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.284320 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.284345 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.284362 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.299311 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.321844 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31375bc51971fa43c64491cb20e8a9466ad12d0ac95044df0eab02849a1d02f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce24f7d7a1ea4da60d22d65753828056f63997aa58de8660229e32aa00a1b9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.335997 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qxrgm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0cd396c-f997-4a5c-b5ec-a901d099bc63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fff5068d1beb0d5a5ffd5ea976fa28f4da9353f6aa835374442927aa3342ecee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpnbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qxrgm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.351501 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b3ef69-0fe7-4e09-9c2d-6c9c95e25540\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fbd3072ee050c9a83353949c6a2f3d761019bd054ae54a97ea97a9aea27fd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://315397d5f35971f9d9b0feba14bf390031b4a146939bc9c5d942fead6f44226f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.368635 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bc33a1a-7cae-4be1-9348-d3f9ebbe84cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a75b9ed748afb4c06594f877d7d715aca537453410c8595fba2eab9229634e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb94f9470852ebca02c2fd758a6b6850e59b429eca8975c46092701755a4cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpkpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sz99b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.384214 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4546f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7721914f-1dfe-4ddc-a0d8-3b9b42de6440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ec1f9170abce2a7cf2aee7339f08bf66d47073dc84c47cf56bbb560c16a6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm22h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4546f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.388677 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.388810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.388896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.388967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.389028 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.403118 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8be6c92a-aa1d-4749-963b-49e360f63b97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b5ed6206a734f3110f7e15fffd0bb0a49eec600eea5dff2bc78f7248dcb52c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kp47b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.421580 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e2bc778-65bf-4304-a4e2-1e31410614c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4742574b68f5c5fd8822d99c1b42d9b87ed7fcc1917ed4bf9028b83046f24d58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab09550404b9b467fc3e7e5d2a09aa07a407ed6f301747a00ee6971dc5e3542\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f0ce949f895a2af3e4a8f285a5ade1349df59512f308642900e0d7c77d34a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://946a292d8e8a8d5ad4a59b582f0f47e9f8e854664a56af3d9f041fb3639a3754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a8f051049af5d3e6319bccabe2bd14f3fcf1e1591c37937b65b35df109776\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e41741aa104d2231b640e62e326eae4b1c182ea1e17fb4b9bfdb71dc31bac663\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1bc7172d2a4d3dde5f5a621035cd6e7709412d9b2a0113882998c0ecb74bf5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwsjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll6w2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.437053 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-strs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"439dd590-d737-435b-80a7-01f75fe6b64d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpl97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-strs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.453501 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b258df0a-140d-401a-abda-48af3552910d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb65bfa5c87a65e6bd1d01344f1ae4bcc7448a6cb2f6d7efc1889116fefc3ab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24139d247f15c76032e81b4a66c134031375aed97e1d90688050f0f0ab956b4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676df03018e370e9644720e41b5a87f6ddfde00ec3ddea0f0ca268881f29e2ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.471954 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b080fdf8-66bf-4676-828b-ae5d239b7c87\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c32250faf0406cb71f279d317c7f5a98a194d666e97e866e222804436590e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51fde96faecd72372b1a6818cabb42ebd77ff4f05c95660b59dcaa6f69c34548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff09354d5d2c1643f1ec9d16aa1694b0f552b7ec88f034ffb8c2334fd966fa94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e470cc46292403f8ea5e9a3a39194c0385fe16c11da7b99f971acad085834338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.492806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.492879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.492898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.492927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.492947 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.493665 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.515616 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75799139375e31b64192df4441e726b6c2f1392dacd66e9b09974989486e201f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.539700 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a113dccf-6146-481e-947f-e782eb9c7b53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:23:17Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1208 19:23:17.629774 6753 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1208 19:23:17.629803 6753 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1208 19:23:17.629817 6753 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1208 19:23:17.629864 6753 factory.go:1336] Added *v1.Node event handler 7\\\\nI1208 19:23:17.629913 6753 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1208 19:23:17.630229 6753 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1208 19:23:17.630321 6753 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1208 19:23:17.630352 6753 ovnkube.go:599] Stopped ovnkube\\\\nI1208 19:23:17.630377 6753 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1208 19:23:17.630448 6753 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:23:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:22:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bptfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kl95h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.562429 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"504b4535-ae3b-440f-acc3-2729c2f9713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T19:22:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 19:22:10.290362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 19:22:10.291715 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3237118473/tls.crt::/tmp/serving-cert-3237118473/tls.key\\\\\\\"\\\\nI1208 19:22:15.662545 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 19:22:15.665689 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 19:22:15.665716 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 19:22:15.665741 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 19:22:15.665748 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 19:22:15.673023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 19:22:15.673070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673078 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 19:22:15.673084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 19:22:15.673088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 19:22:15.673093 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 19:22:15.673096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1208 19:22:15.673293 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1208 19:22:15.677464 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:21:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T19:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T19:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:21:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.594076 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea8020a87e010a43a8cebfd6934aaaa830747d95ee7b9751f8bcb2c1ca05796c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:22:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.596538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.596594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.596643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.596662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.596673 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.607658 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:19 crc kubenswrapper[4706]: E1208 19:23:19.608202 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.616244 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.638390 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x674x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31d41aa9-3a74-42f4-9933-3f484a89c41d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:22:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T19:23:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T19:23:07Z\\\",\\\"message\\\":\\\"2025-12-08T19:22:22+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3\\\\n2025-12-08T19:22:22+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e844e555-cd17-4eae-b666-bb5230a31aa3 to /host/opt/cni/bin/\\\\n2025-12-08T19:22:22Z [verbose] multus-daemon started\\\\n2025-12-08T19:22:22Z [verbose] Readiness Indicator file check\\\\n2025-12-08T19:23:07Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T19:22:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T19:23:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9vhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T19:22:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x674x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T19:23:19Z is after 2025-08-24T17:21:41Z" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.699949 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.700081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.700103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.700137 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.700157 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.803664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.803737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.803754 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.803806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.803834 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.907196 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.907257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.907283 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.907300 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:19 crc kubenswrapper[4706]: I1208 19:23:19.907311 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:19Z","lastTransitionTime":"2025-12-08T19:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.011100 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.011169 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.011187 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.011220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.011239 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:20Z","lastTransitionTime":"2025-12-08T19:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.114367 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.114439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.114481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.114524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.114550 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:20Z","lastTransitionTime":"2025-12-08T19:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.166861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.166932 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.166954 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.166988 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.167010 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T19:23:20Z","lastTransitionTime":"2025-12-08T19:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.247400 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw"] Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.248199 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.251414 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.253472 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.255842 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.261787 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.307940 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=64.307906172 podStartE2EDuration="1m4.307906172s" podCreationTimestamp="2025-12-08 19:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.301042227 +0000 UTC m=+82.943243290" watchObservedRunningTime="2025-12-08 19:23:20.307906172 +0000 UTC m=+82.950107215" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.387967 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-qxrgm" podStartSLOduration=60.387934464 podStartE2EDuration="1m0.387934464s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.368941346 +0000 UTC m=+83.011142389" watchObservedRunningTime="2025-12-08 19:23:20.387934464 +0000 UTC m=+83.030135507" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.409938 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sz99b" podStartSLOduration=59.409908221 podStartE2EDuration="59.409908221s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.409237821 +0000 UTC m=+83.051438864" watchObservedRunningTime="2025-12-08 19:23:20.409908221 +0000 UTC m=+83.052109264" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.410325 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.410316043 podStartE2EDuration="10.410316043s" podCreationTimestamp="2025-12-08 19:23:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.389182021 +0000 UTC m=+83.031383064" watchObservedRunningTime="2025-12-08 19:23:20.410316043 +0000 UTC m=+83.052517076" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.411507 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.411575 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.411625 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.411671 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.411869 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.443501 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=65.443464743 podStartE2EDuration="1m5.443464743s" podCreationTimestamp="2025-12-08 19:22:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.442424232 +0000 UTC m=+83.084625295" watchObservedRunningTime="2025-12-08 19:23:20.443464743 +0000 UTC m=+83.085665776" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.473873 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=31.473849622 podStartE2EDuration="31.473849622s" podCreationTimestamp="2025-12-08 19:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.472742048 +0000 UTC m=+83.114943051" watchObservedRunningTime="2025-12-08 19:23:20.473849622 +0000 UTC m=+83.116050635" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513137 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.513320 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:24.5132974 +0000 UTC m=+147.155498413 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513360 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513403 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513428 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513450 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513472 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513496 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513518 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513546 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513575 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.513644 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.513680 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.513690 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.513706 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.513759 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:24:24.513737574 +0000 UTC m=+147.155938587 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.513792 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 19:24:24.513781255 +0000 UTC m=+147.155982268 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513735 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.513793 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.514112 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.514207 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 19:24:24.514184347 +0000 UTC m=+147.156385350 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.514454 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.514490 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.514512 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.514577 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 19:24:24.514560568 +0000 UTC m=+147.156761751 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.514632 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.528218 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.537142 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8fe45209-2984-4b0a-98b9-5d0dd6ea1d08-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hfvsw\" (UID: \"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.542499 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-4546f" podStartSLOduration=60.542472952 podStartE2EDuration="1m0.542472952s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.54204653 +0000 UTC m=+83.184247543" watchObservedRunningTime="2025-12-08 19:23:20.542472952 +0000 UTC m=+83.184673965" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.555966 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podStartSLOduration=60.555941065 podStartE2EDuration="1m0.555941065s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.555315306 +0000 UTC m=+83.197516309" watchObservedRunningTime="2025-12-08 19:23:20.555941065 +0000 UTC m=+83.198142068" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.572219 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.579614 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ll6w2" podStartSLOduration=60.579584232 podStartE2EDuration="1m0.579584232s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.578960063 +0000 UTC m=+83.221161066" watchObservedRunningTime="2025-12-08 19:23:20.579584232 +0000 UTC m=+83.221785245" Dec 08 19:23:20 crc kubenswrapper[4706]: W1208 19:23:20.586318 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fe45209_2984_4b0a_98b9_5d0dd6ea1d08.slice/crio-33748704f21576e1ef9db90604b7f8b0f02f955d32cebd983da374bbb5293cc3 WatchSource:0}: Error finding container 33748704f21576e1ef9db90604b7f8b0f02f955d32cebd983da374bbb5293cc3: Status 404 returned error can't find the container with id 33748704f21576e1ef9db90604b7f8b0f02f955d32cebd983da374bbb5293cc3 Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.608283 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.608688 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.608987 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.609070 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.609243 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:20 crc kubenswrapper[4706]: E1208 19:23:20.609339 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.625564 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=65.625544085 podStartE2EDuration="1m5.625544085s" podCreationTimestamp="2025-12-08 19:22:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.623599307 +0000 UTC m=+83.265800320" watchObservedRunningTime="2025-12-08 19:23:20.625544085 +0000 UTC m=+83.267745098" Dec 08 19:23:20 crc kubenswrapper[4706]: I1208 19:23:20.685233 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-x674x" podStartSLOduration=60.685213178 podStartE2EDuration="1m0.685213178s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:20.685097245 +0000 UTC m=+83.327298278" watchObservedRunningTime="2025-12-08 19:23:20.685213178 +0000 UTC m=+83.327414181" Dec 08 19:23:21 crc kubenswrapper[4706]: I1208 19:23:21.247859 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" event={"ID":"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08","Type":"ContainerStarted","Data":"2079ba103fefcbf885423963e3986a14f26bd7f16f3249c49bcee4f64dd1a812"} Dec 08 19:23:21 crc kubenswrapper[4706]: I1208 19:23:21.247910 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" event={"ID":"8fe45209-2984-4b0a-98b9-5d0dd6ea1d08","Type":"ContainerStarted","Data":"33748704f21576e1ef9db90604b7f8b0f02f955d32cebd983da374bbb5293cc3"} Dec 08 19:23:21 crc kubenswrapper[4706]: I1208 19:23:21.269457 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hfvsw" podStartSLOduration=61.269435309 podStartE2EDuration="1m1.269435309s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:23:21.268785739 +0000 UTC m=+83.910986742" watchObservedRunningTime="2025-12-08 19:23:21.269435309 +0000 UTC m=+83.911636312" Dec 08 19:23:21 crc kubenswrapper[4706]: I1208 19:23:21.607761 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:21 crc kubenswrapper[4706]: E1208 19:23:21.607954 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:22 crc kubenswrapper[4706]: I1208 19:23:22.608025 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:22 crc kubenswrapper[4706]: I1208 19:23:22.608168 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:22 crc kubenswrapper[4706]: I1208 19:23:22.608183 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:22 crc kubenswrapper[4706]: E1208 19:23:22.608405 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:22 crc kubenswrapper[4706]: E1208 19:23:22.608660 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:22 crc kubenswrapper[4706]: E1208 19:23:22.608773 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:23 crc kubenswrapper[4706]: I1208 19:23:23.607950 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:23 crc kubenswrapper[4706]: E1208 19:23:23.608101 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:24 crc kubenswrapper[4706]: I1208 19:23:24.608167 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:24 crc kubenswrapper[4706]: E1208 19:23:24.608775 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:24 crc kubenswrapper[4706]: I1208 19:23:24.608233 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:24 crc kubenswrapper[4706]: I1208 19:23:24.608171 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:24 crc kubenswrapper[4706]: E1208 19:23:24.608863 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:24 crc kubenswrapper[4706]: E1208 19:23:24.609004 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:25 crc kubenswrapper[4706]: I1208 19:23:25.608611 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:25 crc kubenswrapper[4706]: E1208 19:23:25.608839 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:26 crc kubenswrapper[4706]: I1208 19:23:26.607876 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:26 crc kubenswrapper[4706]: E1208 19:23:26.608439 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:26 crc kubenswrapper[4706]: I1208 19:23:26.607959 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:26 crc kubenswrapper[4706]: I1208 19:23:26.607925 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:26 crc kubenswrapper[4706]: E1208 19:23:26.608819 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:26 crc kubenswrapper[4706]: E1208 19:23:26.608715 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:27 crc kubenswrapper[4706]: I1208 19:23:27.607471 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:27 crc kubenswrapper[4706]: E1208 19:23:27.610975 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:28 crc kubenswrapper[4706]: I1208 19:23:28.607209 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:28 crc kubenswrapper[4706]: I1208 19:23:28.607220 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:28 crc kubenswrapper[4706]: E1208 19:23:28.607345 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:28 crc kubenswrapper[4706]: I1208 19:23:28.607305 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:28 crc kubenswrapper[4706]: E1208 19:23:28.607406 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:28 crc kubenswrapper[4706]: E1208 19:23:28.612914 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:29 crc kubenswrapper[4706]: I1208 19:23:29.607891 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:29 crc kubenswrapper[4706]: E1208 19:23:29.608246 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:30 crc kubenswrapper[4706]: I1208 19:23:30.607923 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:30 crc kubenswrapper[4706]: I1208 19:23:30.607991 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:30 crc kubenswrapper[4706]: I1208 19:23:30.608108 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:30 crc kubenswrapper[4706]: E1208 19:23:30.608187 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:30 crc kubenswrapper[4706]: E1208 19:23:30.608431 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:30 crc kubenswrapper[4706]: E1208 19:23:30.608525 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:31 crc kubenswrapper[4706]: I1208 19:23:31.608943 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:31 crc kubenswrapper[4706]: E1208 19:23:31.609117 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:31 crc kubenswrapper[4706]: I1208 19:23:31.609453 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:23:31 crc kubenswrapper[4706]: E1208 19:23:31.609859 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:23:32 crc kubenswrapper[4706]: I1208 19:23:32.608282 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:32 crc kubenswrapper[4706]: I1208 19:23:32.608289 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:32 crc kubenswrapper[4706]: I1208 19:23:32.608680 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:32 crc kubenswrapper[4706]: E1208 19:23:32.608975 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:32 crc kubenswrapper[4706]: E1208 19:23:32.609190 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:32 crc kubenswrapper[4706]: E1208 19:23:32.609343 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:33 crc kubenswrapper[4706]: I1208 19:23:33.608215 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:33 crc kubenswrapper[4706]: E1208 19:23:33.608759 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:34 crc kubenswrapper[4706]: I1208 19:23:34.607237 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:34 crc kubenswrapper[4706]: I1208 19:23:34.607295 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:34 crc kubenswrapper[4706]: E1208 19:23:34.607380 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:34 crc kubenswrapper[4706]: I1208 19:23:34.607312 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:34 crc kubenswrapper[4706]: E1208 19:23:34.607446 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:34 crc kubenswrapper[4706]: E1208 19:23:34.607565 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:35 crc kubenswrapper[4706]: I1208 19:23:35.608073 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:35 crc kubenswrapper[4706]: E1208 19:23:35.608323 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:36 crc kubenswrapper[4706]: I1208 19:23:36.607772 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:36 crc kubenswrapper[4706]: I1208 19:23:36.607849 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:36 crc kubenswrapper[4706]: I1208 19:23:36.607972 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:36 crc kubenswrapper[4706]: E1208 19:23:36.608080 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:36 crc kubenswrapper[4706]: E1208 19:23:36.608514 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:36 crc kubenswrapper[4706]: E1208 19:23:36.608568 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:37 crc kubenswrapper[4706]: I1208 19:23:37.607678 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:37 crc kubenswrapper[4706]: E1208 19:23:37.607894 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:38 crc kubenswrapper[4706]: I1208 19:23:38.607393 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:38 crc kubenswrapper[4706]: I1208 19:23:38.607485 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:38 crc kubenswrapper[4706]: I1208 19:23:38.607524 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:38 crc kubenswrapper[4706]: E1208 19:23:38.607810 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:38 crc kubenswrapper[4706]: E1208 19:23:38.607918 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:38 crc kubenswrapper[4706]: E1208 19:23:38.608418 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:39 crc kubenswrapper[4706]: I1208 19:23:39.532521 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:39 crc kubenswrapper[4706]: E1208 19:23:39.532702 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:23:39 crc kubenswrapper[4706]: E1208 19:23:39.532799 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs podName:439dd590-d737-435b-80a7-01f75fe6b64d nodeName:}" failed. No retries permitted until 2025-12-08 19:24:43.532778627 +0000 UTC m=+166.174979640 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs") pod "network-metrics-daemon-strs6" (UID: "439dd590-d737-435b-80a7-01f75fe6b64d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 19:23:39 crc kubenswrapper[4706]: I1208 19:23:39.607978 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:39 crc kubenswrapper[4706]: E1208 19:23:39.608362 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:40 crc kubenswrapper[4706]: I1208 19:23:40.607409 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:40 crc kubenswrapper[4706]: I1208 19:23:40.607454 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:40 crc kubenswrapper[4706]: I1208 19:23:40.607526 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:40 crc kubenswrapper[4706]: E1208 19:23:40.607877 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:40 crc kubenswrapper[4706]: E1208 19:23:40.607968 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:40 crc kubenswrapper[4706]: E1208 19:23:40.608068 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:41 crc kubenswrapper[4706]: I1208 19:23:41.607980 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:41 crc kubenswrapper[4706]: E1208 19:23:41.608192 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:42 crc kubenswrapper[4706]: I1208 19:23:42.607564 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:42 crc kubenswrapper[4706]: I1208 19:23:42.607649 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:42 crc kubenswrapper[4706]: I1208 19:23:42.607575 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:42 crc kubenswrapper[4706]: E1208 19:23:42.607763 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:42 crc kubenswrapper[4706]: E1208 19:23:42.607938 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:42 crc kubenswrapper[4706]: E1208 19:23:42.608037 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:43 crc kubenswrapper[4706]: I1208 19:23:43.608073 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:43 crc kubenswrapper[4706]: E1208 19:23:43.608741 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:44 crc kubenswrapper[4706]: I1208 19:23:44.608205 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:44 crc kubenswrapper[4706]: I1208 19:23:44.608582 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:44 crc kubenswrapper[4706]: I1208 19:23:44.608724 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:44 crc kubenswrapper[4706]: E1208 19:23:44.608970 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:44 crc kubenswrapper[4706]: I1208 19:23:44.609119 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:23:44 crc kubenswrapper[4706]: E1208 19:23:44.609209 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:44 crc kubenswrapper[4706]: E1208 19:23:44.609116 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:44 crc kubenswrapper[4706]: E1208 19:23:44.609365 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:23:45 crc kubenswrapper[4706]: I1208 19:23:45.607961 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:45 crc kubenswrapper[4706]: E1208 19:23:45.608122 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:46 crc kubenswrapper[4706]: I1208 19:23:46.608133 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:46 crc kubenswrapper[4706]: I1208 19:23:46.608195 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:46 crc kubenswrapper[4706]: I1208 19:23:46.608133 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:46 crc kubenswrapper[4706]: E1208 19:23:46.608438 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:46 crc kubenswrapper[4706]: E1208 19:23:46.608568 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:46 crc kubenswrapper[4706]: E1208 19:23:46.608849 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:47 crc kubenswrapper[4706]: I1208 19:23:47.608249 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:47 crc kubenswrapper[4706]: E1208 19:23:47.610885 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:48 crc kubenswrapper[4706]: I1208 19:23:48.607768 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:48 crc kubenswrapper[4706]: I1208 19:23:48.607811 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:48 crc kubenswrapper[4706]: I1208 19:23:48.607794 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:48 crc kubenswrapper[4706]: E1208 19:23:48.608005 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:48 crc kubenswrapper[4706]: E1208 19:23:48.608139 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:48 crc kubenswrapper[4706]: E1208 19:23:48.608396 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:49 crc kubenswrapper[4706]: I1208 19:23:49.608061 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:49 crc kubenswrapper[4706]: E1208 19:23:49.608210 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:50 crc kubenswrapper[4706]: I1208 19:23:50.607731 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:50 crc kubenswrapper[4706]: I1208 19:23:50.607820 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:50 crc kubenswrapper[4706]: I1208 19:23:50.607761 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:50 crc kubenswrapper[4706]: E1208 19:23:50.608001 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:50 crc kubenswrapper[4706]: E1208 19:23:50.608125 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:50 crc kubenswrapper[4706]: E1208 19:23:50.608254 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:51 crc kubenswrapper[4706]: I1208 19:23:51.607936 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:51 crc kubenswrapper[4706]: E1208 19:23:51.608174 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:52 crc kubenswrapper[4706]: I1208 19:23:52.607310 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:52 crc kubenswrapper[4706]: I1208 19:23:52.607439 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:52 crc kubenswrapper[4706]: I1208 19:23:52.607538 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:52 crc kubenswrapper[4706]: E1208 19:23:52.607722 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:52 crc kubenswrapper[4706]: E1208 19:23:52.607898 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:52 crc kubenswrapper[4706]: E1208 19:23:52.608133 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:53 crc kubenswrapper[4706]: I1208 19:23:53.608084 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:53 crc kubenswrapper[4706]: E1208 19:23:53.608239 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.368407 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/1.log" Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.369369 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/0.log" Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.369421 4706 generic.go:334] "Generic (PLEG): container finished" podID="31d41aa9-3a74-42f4-9933-3f484a89c41d" containerID="f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22" exitCode=1 Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.369459 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x674x" event={"ID":"31d41aa9-3a74-42f4-9933-3f484a89c41d","Type":"ContainerDied","Data":"f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22"} Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.369500 4706 scope.go:117] "RemoveContainer" containerID="92b80ff932b020cb43accddb21dc00470bef2400a104982f21eac5013c19cba5" Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.369997 4706 scope.go:117] "RemoveContainer" containerID="f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22" Dec 08 19:23:54 crc kubenswrapper[4706]: E1208 19:23:54.370184 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-x674x_openshift-multus(31d41aa9-3a74-42f4-9933-3f484a89c41d)\"" pod="openshift-multus/multus-x674x" podUID="31d41aa9-3a74-42f4-9933-3f484a89c41d" Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.607971 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.608009 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:54 crc kubenswrapper[4706]: I1208 19:23:54.607988 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:54 crc kubenswrapper[4706]: E1208 19:23:54.608146 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:54 crc kubenswrapper[4706]: E1208 19:23:54.608285 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:54 crc kubenswrapper[4706]: E1208 19:23:54.608470 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:55 crc kubenswrapper[4706]: I1208 19:23:55.374748 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/1.log" Dec 08 19:23:55 crc kubenswrapper[4706]: I1208 19:23:55.608613 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:55 crc kubenswrapper[4706]: E1208 19:23:55.608844 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:56 crc kubenswrapper[4706]: I1208 19:23:56.607815 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:56 crc kubenswrapper[4706]: I1208 19:23:56.607815 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:56 crc kubenswrapper[4706]: E1208 19:23:56.608040 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:56 crc kubenswrapper[4706]: I1208 19:23:56.608006 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:56 crc kubenswrapper[4706]: E1208 19:23:56.608190 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:56 crc kubenswrapper[4706]: E1208 19:23:56.608482 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:57 crc kubenswrapper[4706]: E1208 19:23:57.594647 4706 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 08 19:23:57 crc kubenswrapper[4706]: I1208 19:23:57.608420 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:57 crc kubenswrapper[4706]: E1208 19:23:57.609331 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:23:57 crc kubenswrapper[4706]: I1208 19:23:57.610141 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:23:57 crc kubenswrapper[4706]: E1208 19:23:57.610373 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kl95h_openshift-ovn-kubernetes(a113dccf-6146-481e-947f-e782eb9c7b53)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" Dec 08 19:23:57 crc kubenswrapper[4706]: E1208 19:23:57.819231 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:23:58 crc kubenswrapper[4706]: I1208 19:23:58.608236 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:23:58 crc kubenswrapper[4706]: I1208 19:23:58.608332 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:23:58 crc kubenswrapper[4706]: I1208 19:23:58.608424 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:23:58 crc kubenswrapper[4706]: E1208 19:23:58.608515 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:23:58 crc kubenswrapper[4706]: E1208 19:23:58.608680 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:23:58 crc kubenswrapper[4706]: E1208 19:23:58.608734 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:23:59 crc kubenswrapper[4706]: I1208 19:23:59.608040 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:23:59 crc kubenswrapper[4706]: E1208 19:23:59.608222 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:24:00 crc kubenswrapper[4706]: I1208 19:24:00.607636 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:00 crc kubenswrapper[4706]: I1208 19:24:00.607834 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:00 crc kubenswrapper[4706]: I1208 19:24:00.608047 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:00 crc kubenswrapper[4706]: E1208 19:24:00.608061 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:24:00 crc kubenswrapper[4706]: E1208 19:24:00.608102 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:24:00 crc kubenswrapper[4706]: E1208 19:24:00.607843 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:24:01 crc kubenswrapper[4706]: I1208 19:24:01.608239 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:01 crc kubenswrapper[4706]: E1208 19:24:01.608515 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:24:02 crc kubenswrapper[4706]: I1208 19:24:02.607523 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:02 crc kubenswrapper[4706]: I1208 19:24:02.607547 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:02 crc kubenswrapper[4706]: E1208 19:24:02.607749 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:24:02 crc kubenswrapper[4706]: E1208 19:24:02.607850 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:24:02 crc kubenswrapper[4706]: I1208 19:24:02.607622 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:02 crc kubenswrapper[4706]: E1208 19:24:02.608041 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:24:02 crc kubenswrapper[4706]: E1208 19:24:02.821302 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:24:03 crc kubenswrapper[4706]: I1208 19:24:03.607869 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:03 crc kubenswrapper[4706]: E1208 19:24:03.608102 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:24:04 crc kubenswrapper[4706]: I1208 19:24:04.608032 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:04 crc kubenswrapper[4706]: I1208 19:24:04.608076 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:04 crc kubenswrapper[4706]: I1208 19:24:04.608033 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:04 crc kubenswrapper[4706]: E1208 19:24:04.608293 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:24:04 crc kubenswrapper[4706]: E1208 19:24:04.608491 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:24:04 crc kubenswrapper[4706]: E1208 19:24:04.608622 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:24:05 crc kubenswrapper[4706]: I1208 19:24:05.607738 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:05 crc kubenswrapper[4706]: E1208 19:24:05.607905 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:24:06 crc kubenswrapper[4706]: I1208 19:24:06.607473 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:06 crc kubenswrapper[4706]: I1208 19:24:06.607524 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:06 crc kubenswrapper[4706]: E1208 19:24:06.607641 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:24:06 crc kubenswrapper[4706]: I1208 19:24:06.607580 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:06 crc kubenswrapper[4706]: E1208 19:24:06.607807 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:24:06 crc kubenswrapper[4706]: E1208 19:24:06.607953 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:24:07 crc kubenswrapper[4706]: I1208 19:24:07.607653 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:07 crc kubenswrapper[4706]: E1208 19:24:07.608596 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:24:07 crc kubenswrapper[4706]: E1208 19:24:07.822533 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:24:08 crc kubenswrapper[4706]: I1208 19:24:08.607842 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:08 crc kubenswrapper[4706]: I1208 19:24:08.608535 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:24:08 crc kubenswrapper[4706]: I1208 19:24:08.608040 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:08 crc kubenswrapper[4706]: E1208 19:24:08.608562 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:24:08 crc kubenswrapper[4706]: I1208 19:24:08.607908 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:08 crc kubenswrapper[4706]: I1208 19:24:08.608743 4706 scope.go:117] "RemoveContainer" containerID="f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22" Dec 08 19:24:08 crc kubenswrapper[4706]: E1208 19:24:08.609160 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:24:08 crc kubenswrapper[4706]: E1208 19:24:08.608862 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.372619 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-strs6"] Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.433324 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/1.log" Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.433451 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x674x" event={"ID":"31d41aa9-3a74-42f4-9933-3f484a89c41d","Type":"ContainerStarted","Data":"17466bfe4afca9ffa7ad363b6a0c5b80401ec79daedb1e59ad77a6f7bc754587"} Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.437175 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/3.log" Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.440111 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:09 crc kubenswrapper[4706]: E1208 19:24:09.440225 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.440595 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerStarted","Data":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.441152 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.488397 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podStartSLOduration=108.488359525 podStartE2EDuration="1m48.488359525s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:09.488061636 +0000 UTC m=+132.130262639" watchObservedRunningTime="2025-12-08 19:24:09.488359525 +0000 UTC m=+132.130560568" Dec 08 19:24:09 crc kubenswrapper[4706]: I1208 19:24:09.607739 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:09 crc kubenswrapper[4706]: E1208 19:24:09.607932 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:24:10 crc kubenswrapper[4706]: I1208 19:24:10.608174 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:10 crc kubenswrapper[4706]: E1208 19:24:10.608697 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:24:10 crc kubenswrapper[4706]: I1208 19:24:10.608314 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:10 crc kubenswrapper[4706]: E1208 19:24:10.608977 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:24:11 crc kubenswrapper[4706]: I1208 19:24:11.607933 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:11 crc kubenswrapper[4706]: I1208 19:24:11.607933 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:11 crc kubenswrapper[4706]: E1208 19:24:11.608249 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 19:24:11 crc kubenswrapper[4706]: E1208 19:24:11.608410 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-strs6" podUID="439dd590-d737-435b-80a7-01f75fe6b64d" Dec 08 19:24:12 crc kubenswrapper[4706]: I1208 19:24:12.607420 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:12 crc kubenswrapper[4706]: I1208 19:24:12.607537 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:12 crc kubenswrapper[4706]: E1208 19:24:12.607632 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 19:24:12 crc kubenswrapper[4706]: E1208 19:24:12.607765 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 19:24:13 crc kubenswrapper[4706]: I1208 19:24:13.608307 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:13 crc kubenswrapper[4706]: I1208 19:24:13.608315 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:13 crc kubenswrapper[4706]: I1208 19:24:13.611115 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 08 19:24:13 crc kubenswrapper[4706]: I1208 19:24:13.611830 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 08 19:24:13 crc kubenswrapper[4706]: I1208 19:24:13.612648 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 08 19:24:13 crc kubenswrapper[4706]: I1208 19:24:13.615815 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 08 19:24:14 crc kubenswrapper[4706]: I1208 19:24:14.607806 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:14 crc kubenswrapper[4706]: I1208 19:24:14.607835 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:14 crc kubenswrapper[4706]: I1208 19:24:14.611671 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 08 19:24:14 crc kubenswrapper[4706]: I1208 19:24:14.611676 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.762099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.827729 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xwgxh"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.828571 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.830473 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.831670 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.832511 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gs7dv"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.833381 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.843365 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.844088 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.845063 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.845063 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.845534 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.845558 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.846054 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.846180 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.846236 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.846323 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.846423 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.846551 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.846595 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.848700 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.849386 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-zgngm"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.849416 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.850382 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.851343 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bxc4t"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.851598 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.851718 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.851604 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.852005 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bxc4t" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.852474 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6rqvg"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.852533 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.852744 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.853346 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.853425 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.853609 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.853848 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.854516 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.854555 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.854564 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.854646 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq5pd"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.854933 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.855120 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.879187 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.879912 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.880704 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.881775 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.883516 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.884117 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.885700 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.884133 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.884460 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.885430 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.886953 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.887280 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.887888 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.888231 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.888510 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.888808 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.889848 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.889970 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.889936 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vh2jf"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.890401 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.892954 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.893150 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.893192 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.893220 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.893523 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.893676 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.894155 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.894310 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.905372 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.905482 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.905882 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.906016 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.906046 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.906129 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.906437 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.906458 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.906725 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.906773 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.907050 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.909197 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.909598 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.910488 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.910772 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.910848 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.910942 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.911077 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.911324 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.911453 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.912091 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.916105 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.920553 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.921098 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.921585 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.921960 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.931518 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-28lq2"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.932819 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j6s78"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.933787 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.934636 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.938338 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.939056 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8d2q2"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.939442 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941427 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941618 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941652 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941676 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-trusted-ca-bundle\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941696 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-encryption-config\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941803 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-serving-cert\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941824 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f5724d1-fcce-4453-9ca0-8d38414cae8f-trusted-ca\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941841 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-encryption-config\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941863 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941883 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr5bn\" (UniqueName: \"kubernetes.io/projected/98220d47-1a06-4e17-90ec-f802f041197c-kube-api-access-kr5bn\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941911 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-service-ca\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941937 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-audit-policies\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941960 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-serving-cert\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941981 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/979bd1cb-70e1-4f2d-b623-aaae40466dff-serving-cert\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.941998 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942015 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98220d47-1a06-4e17-90ec-f802f041197c-serving-cert\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942031 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942046 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-dir\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942061 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-serving-cert\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942077 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/37e229d7-a86c-4e5b-bf51-caec4edc7c39-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lgtpm\" (UID: \"37e229d7-a86c-4e5b-bf51-caec4edc7c39\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942095 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0fd938d-859f-4620-944c-598cbe045a1a-serving-cert\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942109 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-etcd-client\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942127 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-client-ca\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942148 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdnwx\" (UniqueName: \"kubernetes.io/projected/979bd1cb-70e1-4f2d-b623-aaae40466dff-kube-api-access-mdnwx\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942174 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-config\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942191 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942206 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t68v8\" (UniqueName: \"kubernetes.io/projected/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-kube-api-access-t68v8\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942230 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgbw6\" (UniqueName: \"kubernetes.io/projected/3f1e0eea-50be-4dd8-a504-dcdd049ae255-kube-api-access-wgbw6\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942248 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt896\" (UniqueName: \"kubernetes.io/projected/e0fd938d-859f-4620-944c-598cbe045a1a-kube-api-access-nt896\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942281 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-audit\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942300 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-config\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942335 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72qld\" (UniqueName: \"kubernetes.io/projected/4124e67a-ef7b-4984-bb29-29525802aa66-kube-api-access-72qld\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942350 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db465e9d-aadc-41bc-ba53-3fb564d4a541-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942655 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-config\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942673 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-client-ca\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942709 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-etcd-client\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942726 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-image-import-ca\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942750 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4124e67a-ef7b-4984-bb29-29525802aa66-audit-dir\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942778 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942795 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-oauth-config\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.944139 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.945648 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.942811 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f5724d1-fcce-4453-9ca0-8d38414cae8f-config\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947204 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947235 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947283 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-config\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947311 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-etcd-serving-ca\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947333 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947363 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-policies\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947386 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947417 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvlj4\" (UniqueName: \"kubernetes.io/projected/7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb-kube-api-access-jvlj4\") pod \"downloads-7954f5f757-bxc4t\" (UID: \"7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb\") " pod="openshift-console/downloads-7954f5f757-bxc4t" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947440 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947449 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c0d44c3b-0f39-4db6-9553-cf23c89fe209-audit-dir\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947469 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aca2089e-f16a-4bf6-9b7b-f3261a525cc3-metrics-tls\") pod \"dns-operator-744455d44c-vh2jf\" (UID: \"aca2089e-f16a-4bf6-9b7b-f3261a525cc3\") " pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947488 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwl47\" (UniqueName: \"kubernetes.io/projected/0f5724d1-fcce-4453-9ca0-8d38414cae8f-kube-api-access-qwl47\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947508 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxjgz\" (UniqueName: \"kubernetes.io/projected/db465e9d-aadc-41bc-ba53-3fb564d4a541-kube-api-access-xxjgz\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947538 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947557 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr5b6\" (UniqueName: \"kubernetes.io/projected/37e229d7-a86c-4e5b-bf51-caec4edc7c39-kube-api-access-sr5b6\") pod \"cluster-samples-operator-665b6dd947-lgtpm\" (UID: \"37e229d7-a86c-4e5b-bf51-caec4edc7c39\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947576 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e0fd938d-859f-4620-944c-598cbe045a1a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947594 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947614 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-oauth-serving-cert\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947634 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f5724d1-fcce-4453-9ca0-8d38414cae8f-serving-cert\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947658 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7cm6\" (UniqueName: \"kubernetes.io/projected/c0d44c3b-0f39-4db6-9553-cf23c89fe209-kube-api-access-b7cm6\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947683 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c0d44c3b-0f39-4db6-9553-cf23c89fe209-node-pullsecrets\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947701 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf5rg\" (UniqueName: \"kubernetes.io/projected/aca2089e-f16a-4bf6-9b7b-f3261a525cc3-kube-api-access-wf5rg\") pod \"dns-operator-744455d44c-vh2jf\" (UID: \"aca2089e-f16a-4bf6-9b7b-f3261a525cc3\") " pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947726 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.947743 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db465e9d-aadc-41bc-ba53-3fb564d4a541-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.951640 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.979195 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.981739 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.982480 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.983947 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2mptf"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.984485 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.984918 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gv9fs"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.985805 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.986067 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.986861 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.987115 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.987186 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.987687 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.987860 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988011 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988085 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988224 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988430 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988499 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988563 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988612 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988797 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988921 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.988986 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.989281 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.989525 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.990909 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.991048 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.991115 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.987134 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.991348 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.991632 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.991811 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.991841 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.992111 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.992321 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.992349 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.992466 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.992586 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.992643 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.992743 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.992961 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.993132 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.995360 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.995601 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.996018 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.997698 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8"] Dec 08 19:24:21 crc kubenswrapper[4706]: I1208 19:24:21.999301 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.002086 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.007501 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.007713 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.007787 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.008516 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.009955 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.010485 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.008577 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.010822 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.011583 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.011999 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.012539 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.012889 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.013301 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.013802 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-55zvr"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.015338 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.015744 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.015978 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.016551 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.016884 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xwgxh"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.018455 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.019001 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.019011 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.020083 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.020784 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rzswq"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.021880 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.022178 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.023374 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-cvvxf"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.024225 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gs7dv"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.024324 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.025391 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.026670 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.027975 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zgngm"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.029291 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.031436 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.032117 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-f7kfm"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.032648 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.033445 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-f7kfm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.035912 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vh2jf"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.035960 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.040661 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.042156 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.043234 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.044937 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-28lq2"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.046580 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8v5f4"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.047438 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.047742 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-ktnbl"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048386 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/979bd1cb-70e1-4f2d-b623-aaae40466dff-serving-cert\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048419 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-service-ca\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048440 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-audit-policies\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048458 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-serving-cert\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048475 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048499 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx5rb\" (UniqueName: \"kubernetes.io/projected/035c9071-5a5b-4204-810d-1ee403423d01-kube-api-access-gx5rb\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048520 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98220d47-1a06-4e17-90ec-f802f041197c-serving-cert\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048537 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048553 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-dir\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048573 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6889021e-b838-4fb3-b664-92efcdf699a2-secret-volume\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048594 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68qxg\" (UniqueName: \"kubernetes.io/projected/eaed88d5-42bb-478a-842e-1326e77fbc1e-kube-api-access-68qxg\") pod \"multus-admission-controller-857f4d67dd-gv9fs\" (UID: \"eaed88d5-42bb-478a-842e-1326e77fbc1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048614 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0fd938d-859f-4620-944c-598cbe045a1a-serving-cert\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048630 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-etcd-client\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048643 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048651 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-serving-cert\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048673 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/37e229d7-a86c-4e5b-bf51-caec4edc7c39-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lgtpm\" (UID: \"37e229d7-a86c-4e5b-bf51-caec4edc7c39\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048697 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ae483b4-35fe-4790-9523-c5e6b1fd3137-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048717 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-client-ca\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048739 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crpc5\" (UniqueName: \"kubernetes.io/projected/69a6c982-b070-4f4a-8b0c-f841fc147a50-kube-api-access-crpc5\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048767 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdnwx\" (UniqueName: \"kubernetes.io/projected/979bd1cb-70e1-4f2d-b623-aaae40466dff-kube-api-access-mdnwx\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048785 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e775a27-59a7-4c73-a2ae-7598d29b91fc-config\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048805 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048828 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-config\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048846 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t68v8\" (UniqueName: \"kubernetes.io/projected/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-kube-api-access-t68v8\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048866 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48960e60-2160-4afa-98ca-e67603fc9545-images\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048886 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/323ba66c-aab0-4d31-982b-b012b399bc5c-trusted-ca\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048907 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt896\" (UniqueName: \"kubernetes.io/projected/e0fd938d-859f-4620-944c-598cbe045a1a-kube-api-access-nt896\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048936 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgbw6\" (UniqueName: \"kubernetes.io/projected/3f1e0eea-50be-4dd8-a504-dcdd049ae255-kube-api-access-wgbw6\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048953 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-audit\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.048995 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-config\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049014 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049050 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-client-ca\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049069 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72qld\" (UniqueName: \"kubernetes.io/projected/4124e67a-ef7b-4984-bb29-29525802aa66-kube-api-access-72qld\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049087 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db465e9d-aadc-41bc-ba53-3fb564d4a541-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049105 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-config\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049121 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6889021e-b838-4fb3-b664-92efcdf699a2-config-volume\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049139 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z25p7\" (UniqueName: \"kubernetes.io/projected/323ba66c-aab0-4d31-982b-b012b399bc5c-kube-api-access-z25p7\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049157 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b92h\" (UniqueName: \"kubernetes.io/projected/dacda7e8-5813-40a1-bf74-b09eb69c4a40-kube-api-access-4b92h\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049175 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-etcd-client\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049191 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-image-import-ca\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049210 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sfs5\" (UniqueName: \"kubernetes.io/projected/4ae483b4-35fe-4790-9523-c5e6b1fd3137-kube-api-access-7sfs5\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049226 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049246 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacda7e8-5813-40a1-bf74-b09eb69c4a40-serving-cert\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049280 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035c9071-5a5b-4204-810d-1ee403423d01-serving-cert\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049301 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qqbl\" (UniqueName: \"kubernetes.io/projected/48960e60-2160-4afa-98ca-e67603fc9545-kube-api-access-2qqbl\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e775a27-59a7-4c73-a2ae-7598d29b91fc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049344 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4124e67a-ef7b-4984-bb29-29525802aa66-audit-dir\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049362 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wv4z\" (UniqueName: \"kubernetes.io/projected/ef62dd12-f35d-4808-a54b-da8fb928592a-kube-api-access-5wv4z\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049379 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-oauth-config\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049395 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f5724d1-fcce-4453-9ca0-8d38414cae8f-config\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049412 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049428 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049445 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049461 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-config\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049477 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69a6c982-b070-4f4a-8b0c-f841fc147a50-webhook-cert\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049495 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-etcd-serving-ca\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049510 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69a6c982-b070-4f4a-8b0c-f841fc147a50-apiservice-cert\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049526 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-config\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049542 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049561 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-policies\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049580 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049598 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvlj4\" (UniqueName: \"kubernetes.io/projected/7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb-kube-api-access-jvlj4\") pod \"downloads-7954f5f757-bxc4t\" (UID: \"7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb\") " pod="openshift-console/downloads-7954f5f757-bxc4t" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049615 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c0d44c3b-0f39-4db6-9553-cf23c89fe209-audit-dir\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049632 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aca2089e-f16a-4bf6-9b7b-f3261a525cc3-metrics-tls\") pod \"dns-operator-744455d44c-vh2jf\" (UID: \"aca2089e-f16a-4bf6-9b7b-f3261a525cc3\") " pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049649 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef62dd12-f35d-4808-a54b-da8fb928592a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049668 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049693 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwl47\" (UniqueName: \"kubernetes.io/projected/0f5724d1-fcce-4453-9ca0-8d38414cae8f-kube-api-access-qwl47\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049714 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxjgz\" (UniqueName: \"kubernetes.io/projected/db465e9d-aadc-41bc-ba53-3fb564d4a541-kube-api-access-xxjgz\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049733 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae483b4-35fe-4790-9523-c5e6b1fd3137-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049751 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e775a27-59a7-4c73-a2ae-7598d29b91fc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049775 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e0fd938d-859f-4620-944c-598cbe045a1a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049825 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049848 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr5b6\" (UniqueName: \"kubernetes.io/projected/37e229d7-a86c-4e5b-bf51-caec4edc7c39-kube-api-access-sr5b6\") pod \"cluster-samples-operator-665b6dd947-lgtpm\" (UID: \"37e229d7-a86c-4e5b-bf51-caec4edc7c39\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049872 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be5d2ba7-9cfa-4952-a142-e15839642dd9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049892 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef62dd12-f35d-4808-a54b-da8fb928592a-proxy-tls\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049915 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-oauth-serving-cert\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049926 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-client-ca\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049936 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f5724d1-fcce-4453-9ca0-8d38414cae8f-serving-cert\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.049997 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7cm6\" (UniqueName: \"kubernetes.io/projected/c0d44c3b-0f39-4db6-9553-cf23c89fe209-kube-api-access-b7cm6\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050028 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c0d44c3b-0f39-4db6-9553-cf23c89fe209-node-pullsecrets\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050050 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf5rg\" (UniqueName: \"kubernetes.io/projected/aca2089e-f16a-4bf6-9b7b-f3261a525cc3-kube-api-access-wf5rg\") pod \"dns-operator-744455d44c-vh2jf\" (UID: \"aca2089e-f16a-4bf6-9b7b-f3261a525cc3\") " pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050074 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050102 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be5d2ba7-9cfa-4952-a142-e15839642dd9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050123 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be5d2ba7-9cfa-4952-a142-e15839642dd9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050148 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db465e9d-aadc-41bc-ba53-3fb564d4a541-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050170 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035c9071-5a5b-4204-810d-1ee403423d01-config\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050193 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/323ba66c-aab0-4d31-982b-b012b399bc5c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050224 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050289 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050314 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48960e60-2160-4afa-98ca-e67603fc9545-config\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050336 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/eaed88d5-42bb-478a-842e-1326e77fbc1e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gv9fs\" (UID: \"eaed88d5-42bb-478a-842e-1326e77fbc1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050372 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-trusted-ca-bundle\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050391 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-encryption-config\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050419 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/323ba66c-aab0-4d31-982b-b012b399bc5c-metrics-tls\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050437 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sj9p\" (UniqueName: \"kubernetes.io/projected/6889021e-b838-4fb3-b664-92efcdf699a2-kube-api-access-9sj9p\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050458 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-serving-cert\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050476 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f5724d1-fcce-4453-9ca0-8d38414cae8f-trusted-ca\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050496 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-encryption-config\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050517 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050538 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr5bn\" (UniqueName: \"kubernetes.io/projected/98220d47-1a06-4e17-90ec-f802f041197c-kube-api-access-kr5bn\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050559 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/48960e60-2160-4afa-98ca-e67603fc9545-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050577 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69a6c982-b070-4f4a-8b0c-f841fc147a50-tmpfs\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.050595 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-service-ca-bundle\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.051143 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.051799 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.051845 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-dir\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.052281 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.052345 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq5pd"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.052357 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.052544 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-config\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.053407 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-audit\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.054430 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.056775 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-config\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.056880 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-client-ca\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.057404 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db465e9d-aadc-41bc-ba53-3fb564d4a541-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.060353 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c0d44c3b-0f39-4db6-9553-cf23c89fe209-node-pullsecrets\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.060603 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.061097 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-trusted-ca-bundle\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.061921 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-audit-policies\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.062619 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.062729 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/979bd1cb-70e1-4f2d-b623-aaae40466dff-serving-cert\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.062922 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.063224 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db465e9d-aadc-41bc-ba53-3fb564d4a541-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.063294 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.063459 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-etcd-serving-ca\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.063997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98220d47-1a06-4e17-90ec-f802f041197c-serving-cert\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.064033 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-service-ca\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.064329 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-oauth-serving-cert\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.065395 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-config\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.065396 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f5724d1-fcce-4453-9ca0-8d38414cae8f-config\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.066239 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c0d44c3b-0f39-4db6-9553-cf23c89fe209-image-import-ca\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.066624 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4124e67a-ef7b-4984-bb29-29525802aa66-audit-dir\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.066661 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bxc4t"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.067002 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-policies\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.067497 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.067942 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-encryption-config\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.068247 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-config\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.069091 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c0d44c3b-0f39-4db6-9553-cf23c89fe209-audit-dir\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.069463 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0fd938d-859f-4620-944c-598cbe045a1a-serving-cert\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.070751 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f5724d1-fcce-4453-9ca0-8d38414cae8f-trusted-ca\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.071298 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.071556 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aca2089e-f16a-4bf6-9b7b-f3261a525cc3-metrics-tls\") pod \"dns-operator-744455d44c-vh2jf\" (UID: \"aca2089e-f16a-4bf6-9b7b-f3261a525cc3\") " pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.071600 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.072522 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e0fd938d-859f-4620-944c-598cbe045a1a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.073985 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.074070 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-oauth-config\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.074235 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4124e67a-ef7b-4984-bb29-29525802aa66-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.074462 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-etcd-client\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.074998 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.075178 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.076403 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.076700 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.079300 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.079405 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.080905 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/37e229d7-a86c-4e5b-bf51-caec4edc7c39-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lgtpm\" (UID: \"37e229d7-a86c-4e5b-bf51-caec4edc7c39\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.082213 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.083324 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.085044 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.086276 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8d2q2"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.087715 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-serving-cert\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.088087 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0d44c3b-0f39-4db6-9553-cf23c89fe209-serving-cert\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.088719 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f5724d1-fcce-4453-9ca0-8d38414cae8f-serving-cert\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.089275 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.090963 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.093410 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6rqvg"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.093848 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-encryption-config\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.094753 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-etcd-client\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.096211 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gv9fs"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.097400 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4124e67a-ef7b-4984-bb29-29525802aa66-serving-cert\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.099006 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2mptf"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.101583 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.103049 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j6s78"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.104601 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.106700 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.108222 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.109380 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.110545 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ktnbl"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.110755 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.111631 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.112755 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.113804 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-f7kfm"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.114828 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-55zvr"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.115909 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rzswq"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.117174 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-q9lfk"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.119174 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-q9lfk"] Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.119291 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.131074 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151409 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151716 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/323ba66c-aab0-4d31-982b-b012b399bc5c-metrics-tls\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151758 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sj9p\" (UniqueName: \"kubernetes.io/projected/6889021e-b838-4fb3-b664-92efcdf699a2-kube-api-access-9sj9p\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151786 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-service-ca-bundle\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/48960e60-2160-4afa-98ca-e67603fc9545-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151836 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69a6c982-b070-4f4a-8b0c-f841fc147a50-tmpfs\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151857 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx5rb\" (UniqueName: \"kubernetes.io/projected/035c9071-5a5b-4204-810d-1ee403423d01-kube-api-access-gx5rb\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151879 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6889021e-b838-4fb3-b664-92efcdf699a2-secret-volume\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151900 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68qxg\" (UniqueName: \"kubernetes.io/projected/eaed88d5-42bb-478a-842e-1326e77fbc1e-kube-api-access-68qxg\") pod \"multus-admission-controller-857f4d67dd-gv9fs\" (UID: \"eaed88d5-42bb-478a-842e-1326e77fbc1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151923 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ae483b4-35fe-4790-9523-c5e6b1fd3137-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151965 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crpc5\" (UniqueName: \"kubernetes.io/projected/69a6c982-b070-4f4a-8b0c-f841fc147a50-kube-api-access-crpc5\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.151998 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e775a27-59a7-4c73-a2ae-7598d29b91fc-config\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152055 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48960e60-2160-4afa-98ca-e67603fc9545-images\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152074 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/323ba66c-aab0-4d31-982b-b012b399bc5c-trusted-ca\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152129 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6889021e-b838-4fb3-b664-92efcdf699a2-config-volume\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152149 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z25p7\" (UniqueName: \"kubernetes.io/projected/323ba66c-aab0-4d31-982b-b012b399bc5c-kube-api-access-z25p7\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152167 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b92h\" (UniqueName: \"kubernetes.io/projected/dacda7e8-5813-40a1-bf74-b09eb69c4a40-kube-api-access-4b92h\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152191 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sfs5\" (UniqueName: \"kubernetes.io/projected/4ae483b4-35fe-4790-9523-c5e6b1fd3137-kube-api-access-7sfs5\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152210 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152247 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacda7e8-5813-40a1-bf74-b09eb69c4a40-serving-cert\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152295 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035c9071-5a5b-4204-810d-1ee403423d01-serving-cert\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152325 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qqbl\" (UniqueName: \"kubernetes.io/projected/48960e60-2160-4afa-98ca-e67603fc9545-kube-api-access-2qqbl\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152347 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e775a27-59a7-4c73-a2ae-7598d29b91fc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152368 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wv4z\" (UniqueName: \"kubernetes.io/projected/ef62dd12-f35d-4808-a54b-da8fb928592a-kube-api-access-5wv4z\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152411 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69a6c982-b070-4f4a-8b0c-f841fc147a50-webhook-cert\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152431 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-config\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152457 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69a6c982-b070-4f4a-8b0c-f841fc147a50-apiservice-cert\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152489 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef62dd12-f35d-4808-a54b-da8fb928592a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152506 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e775a27-59a7-4c73-a2ae-7598d29b91fc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152539 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae483b4-35fe-4790-9523-c5e6b1fd3137-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152564 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be5d2ba7-9cfa-4952-a142-e15839642dd9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152587 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef62dd12-f35d-4808-a54b-da8fb928592a-proxy-tls\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152621 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be5d2ba7-9cfa-4952-a142-e15839642dd9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152641 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be5d2ba7-9cfa-4952-a142-e15839642dd9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152661 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035c9071-5a5b-4204-810d-1ee403423d01-config\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152678 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/323ba66c-aab0-4d31-982b-b012b399bc5c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152697 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/eaed88d5-42bb-478a-842e-1326e77fbc1e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gv9fs\" (UID: \"eaed88d5-42bb-478a-842e-1326e77fbc1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.152715 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48960e60-2160-4afa-98ca-e67603fc9545-config\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.153516 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69a6c982-b070-4f4a-8b0c-f841fc147a50-tmpfs\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.153954 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48960e60-2160-4afa-98ca-e67603fc9545-config\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.153977 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae483b4-35fe-4790-9523-c5e6b1fd3137-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.154094 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/323ba66c-aab0-4d31-982b-b012b399bc5c-trusted-ca\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.154197 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef62dd12-f35d-4808-a54b-da8fb928592a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.154776 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48960e60-2160-4afa-98ca-e67603fc9545-images\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.155584 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/323ba66c-aab0-4d31-982b-b012b399bc5c-metrics-tls\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.156812 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/48960e60-2160-4afa-98ca-e67603fc9545-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.157748 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/eaed88d5-42bb-478a-842e-1326e77fbc1e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gv9fs\" (UID: \"eaed88d5-42bb-478a-842e-1326e77fbc1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.160619 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ae483b4-35fe-4790-9523-c5e6b1fd3137-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.171816 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.190644 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.211386 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.230305 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.251712 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.255669 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e775a27-59a7-4c73-a2ae-7598d29b91fc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.271415 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.273107 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e775a27-59a7-4c73-a2ae-7598d29b91fc-config\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.290747 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.310429 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.315837 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacda7e8-5813-40a1-bf74-b09eb69c4a40-serving-cert\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.330838 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.334553 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-config\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.357016 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.367081 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.370866 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.373027 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacda7e8-5813-40a1-bf74-b09eb69c4a40-service-ca-bundle\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.392514 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.411125 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.432104 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.450892 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.470912 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.491823 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.511154 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.531543 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.551775 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.570808 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.591042 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.598343 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be5d2ba7-9cfa-4952-a142-e15839642dd9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.611024 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.614940 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be5d2ba7-9cfa-4952-a142-e15839642dd9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.630933 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.651552 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.656224 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69a6c982-b070-4f4a-8b0c-f841fc147a50-webhook-cert\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.658852 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69a6c982-b070-4f4a-8b0c-f841fc147a50-apiservice-cert\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.671795 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.691976 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.695090 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6889021e-b838-4fb3-b664-92efcdf699a2-config-volume\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.711327 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.731046 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.751368 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.756542 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6889021e-b838-4fb3-b664-92efcdf699a2-secret-volume\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.771584 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.791423 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.811563 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.832693 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.837958 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef62dd12-f35d-4808-a54b-da8fb928592a-proxy-tls\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.851574 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.870878 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.876699 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035c9071-5a5b-4204-810d-1ee403423d01-serving-cert\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.891773 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.911503 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.914671 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035c9071-5a5b-4204-810d-1ee403423d01-config\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.931896 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.951981 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.972149 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 08 19:24:22 crc kubenswrapper[4706]: I1208 19:24:22.991326 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.012591 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.029345 4706 request.go:700] Waited for 1.013309031s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dmarketplace-trusted-ca&limit=500&resourceVersion=0 Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.040703 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.051695 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.072069 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.112088 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.131901 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.151485 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.171401 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.192250 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.211659 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.231799 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.251779 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.272819 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.292703 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.312482 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.331573 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.352402 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.372120 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.391675 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.412771 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.432669 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.452571 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.472069 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.491197 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.532522 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.551886 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.572095 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.591909 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.611603 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.631662 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.652466 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.673023 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.691733 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.712220 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.732682 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.751712 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.772367 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.791176 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.811961 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.831208 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.867594 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdnwx\" (UniqueName: \"kubernetes.io/projected/979bd1cb-70e1-4f2d-b623-aaae40466dff-kube-api-access-mdnwx\") pod \"controller-manager-879f6c89f-xwgxh\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.886382 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t68v8\" (UniqueName: \"kubernetes.io/projected/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-kube-api-access-t68v8\") pod \"oauth-openshift-558db77b4-bq5pd\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.906447 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt896\" (UniqueName: \"kubernetes.io/projected/e0fd938d-859f-4620-944c-598cbe045a1a-kube-api-access-nt896\") pod \"openshift-config-operator-7777fb866f-tkq2v\" (UID: \"e0fd938d-859f-4620-944c-598cbe045a1a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.927134 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.928093 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgbw6\" (UniqueName: \"kubernetes.io/projected/3f1e0eea-50be-4dd8-a504-dcdd049ae255-kube-api-access-wgbw6\") pod \"console-f9d7485db-zgngm\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.951998 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72qld\" (UniqueName: \"kubernetes.io/projected/4124e67a-ef7b-4984-bb29-29525802aa66-kube-api-access-72qld\") pod \"apiserver-7bbb656c7d-t4qw2\" (UID: \"4124e67a-ef7b-4984-bb29-29525802aa66\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.964907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7cm6\" (UniqueName: \"kubernetes.io/projected/c0d44c3b-0f39-4db6-9553-cf23c89fe209-kube-api-access-b7cm6\") pod \"apiserver-76f77b778f-6rqvg\" (UID: \"c0d44c3b-0f39-4db6-9553-cf23c89fe209\") " pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.984057 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf5rg\" (UniqueName: \"kubernetes.io/projected/aca2089e-f16a-4bf6-9b7b-f3261a525cc3-kube-api-access-wf5rg\") pod \"dns-operator-744455d44c-vh2jf\" (UID: \"aca2089e-f16a-4bf6-9b7b-f3261a525cc3\") " pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" Dec 08 19:24:23 crc kubenswrapper[4706]: I1208 19:24:23.985992 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.010962 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxjgz\" (UniqueName: \"kubernetes.io/projected/db465e9d-aadc-41bc-ba53-3fb564d4a541-kube-api-access-xxjgz\") pod \"openshift-controller-manager-operator-756b6f6bc6-2rns4\" (UID: \"db465e9d-aadc-41bc-ba53-3fb564d4a541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.026748 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr5b6\" (UniqueName: \"kubernetes.io/projected/37e229d7-a86c-4e5b-bf51-caec4edc7c39-kube-api-access-sr5b6\") pod \"cluster-samples-operator-665b6dd947-lgtpm\" (UID: \"37e229d7-a86c-4e5b-bf51-caec4edc7c39\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.030142 4706 request.go:700] Waited for 1.962463908s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.030381 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.056926 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwl47\" (UniqueName: \"kubernetes.io/projected/0f5724d1-fcce-4453-9ca0-8d38414cae8f-kube-api-access-qwl47\") pod \"console-operator-58897d9998-gs7dv\" (UID: \"0f5724d1-fcce-4453-9ca0-8d38414cae8f\") " pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.075413 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvlj4\" (UniqueName: \"kubernetes.io/projected/7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb-kube-api-access-jvlj4\") pod \"downloads-7954f5f757-bxc4t\" (UID: \"7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb\") " pod="openshift-console/downloads-7954f5f757-bxc4t" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.086623 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.090946 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr5bn\" (UniqueName: \"kubernetes.io/projected/98220d47-1a06-4e17-90ec-f802f041197c-kube-api-access-kr5bn\") pod \"route-controller-manager-6576b87f9c-pk89q\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.094532 4706 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.110219 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.111497 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.120463 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.129957 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.134379 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.141852 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.153913 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.183245 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sj9p\" (UniqueName: \"kubernetes.io/projected/6889021e-b838-4fb3-b664-92efcdf699a2-kube-api-access-9sj9p\") pod \"collect-profiles-29420355-cpd8r\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.194806 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68qxg\" (UniqueName: \"kubernetes.io/projected/eaed88d5-42bb-478a-842e-1326e77fbc1e-kube-api-access-68qxg\") pod \"multus-admission-controller-857f4d67dd-gv9fs\" (UID: \"eaed88d5-42bb-478a-842e-1326e77fbc1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.209657 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx5rb\" (UniqueName: \"kubernetes.io/projected/035c9071-5a5b-4204-810d-1ee403423d01-kube-api-access-gx5rb\") pod \"service-ca-operator-777779d784-8h7dr\" (UID: \"035c9071-5a5b-4204-810d-1ee403423d01\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.210005 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bxc4t" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.219762 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.235531 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crpc5\" (UniqueName: \"kubernetes.io/projected/69a6c982-b070-4f4a-8b0c-f841fc147a50-kube-api-access-crpc5\") pod \"packageserver-d55dfcdfc-9nzm8\" (UID: \"69a6c982-b070-4f4a-8b0c-f841fc147a50\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.238061 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.252103 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wv4z\" (UniqueName: \"kubernetes.io/projected/ef62dd12-f35d-4808-a54b-da8fb928592a-kube-api-access-5wv4z\") pod \"machine-config-controller-84d6567774-ggc6b\" (UID: \"ef62dd12-f35d-4808-a54b-da8fb928592a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.286320 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qqbl\" (UniqueName: \"kubernetes.io/projected/48960e60-2160-4afa-98ca-e67603fc9545-kube-api-access-2qqbl\") pod \"machine-api-operator-5694c8668f-j6s78\" (UID: \"48960e60-2160-4afa-98ca-e67603fc9545\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.293825 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e775a27-59a7-4c73-a2ae-7598d29b91fc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fddbl\" (UID: \"0e775a27-59a7-4c73-a2ae-7598d29b91fc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.294083 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.309850 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.312728 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be5d2ba7-9cfa-4952-a142-e15839642dd9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sg66n\" (UID: \"be5d2ba7-9cfa-4952-a142-e15839642dd9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.317469 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xwgxh"] Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.328658 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/323ba66c-aab0-4d31-982b-b012b399bc5c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.346662 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b92h\" (UniqueName: \"kubernetes.io/projected/dacda7e8-5813-40a1-bf74-b09eb69c4a40-kube-api-access-4b92h\") pod \"authentication-operator-69f744f599-2mptf\" (UID: \"dacda7e8-5813-40a1-bf74-b09eb69c4a40\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.352960 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" Dec 08 19:24:24 crc kubenswrapper[4706]: W1208 19:24:24.354651 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod979bd1cb_70e1_4f2d_b623_aaae40466dff.slice/crio-26ecfa47e41147d5b5e8d62fab925acb3569a9e751b3db51bb748332c180de07 WatchSource:0}: Error finding container 26ecfa47e41147d5b5e8d62fab925acb3569a9e751b3db51bb748332c180de07: Status 404 returned error can't find the container with id 26ecfa47e41147d5b5e8d62fab925acb3569a9e751b3db51bb748332c180de07 Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.362797 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.366835 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.370699 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sfs5\" (UniqueName: \"kubernetes.io/projected/4ae483b4-35fe-4790-9523-c5e6b1fd3137-kube-api-access-7sfs5\") pod \"openshift-apiserver-operator-796bbdcf4f-qr4hz\" (UID: \"4ae483b4-35fe-4790-9523-c5e6b1fd3137\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.375504 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.388072 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z25p7\" (UniqueName: \"kubernetes.io/projected/323ba66c-aab0-4d31-982b-b012b399bc5c-kube-api-access-z25p7\") pod \"ingress-operator-5b745b69d9-97l9j\" (UID: \"323ba66c-aab0-4d31-982b-b012b399bc5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.399006 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.401925 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq5pd"] Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.453382 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vh2jf"] Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490045 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-config\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490084 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bed16b63-2523-4c5e-88b9-390cac136d1b-profile-collector-cert\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490116 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-bound-sa-token\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490169 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490193 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-tls\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490214 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-trusted-ca\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490338 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/98db57c6-9269-4d83-abd2-d71c34a20f23-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490385 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9n54\" (UniqueName: \"kubernetes.io/projected/bed16b63-2523-4c5e-88b9-390cac136d1b-kube-api-access-q9n54\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490412 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppz7x\" (UniqueName: \"kubernetes.io/projected/8693ec42-f6ae-4cf8-bcde-52843820ecd5-kube-api-access-ppz7x\") pod \"migrator-59844c95c7-gz9d5\" (UID: \"8693ec42-f6ae-4cf8-bcde-52843820ecd5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490436 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82da14d7-a213-4d6c-85f9-3f6f62e48323-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490488 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/25a09c7e-68bf-4a29-9bd5-909468bdcc42-installation-pull-secrets\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490515 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df037075-04d9-4eab-bbc9-86ede7236798-etcd-client\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490571 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/98db57c6-9269-4d83-abd2-d71c34a20f23-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490602 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7112c6d3-009f-4a78-b6c3-5d61eaae0940-srv-cert\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490626 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82da14d7-a213-4d6c-85f9-3f6f62e48323-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490653 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2gz5\" (UniqueName: \"kubernetes.io/projected/d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6-kube-api-access-s2gz5\") pod \"package-server-manager-789f6589d5-n2bsv\" (UID: \"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490693 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bed16b63-2523-4c5e-88b9-390cac136d1b-srv-cert\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490770 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-etcd-service-ca\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490797 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98db57c6-9269-4d83-abd2-d71c34a20f23-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490841 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82da14d7-a213-4d6c-85f9-3f6f62e48323-config\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490865 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490890 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/878ce0cf-8408-4bd3-90cc-948a66892c07-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mxb2z\" (UID: \"878ce0cf-8408-4bd3-90cc-948a66892c07\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490949 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2p9h\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-kube-api-access-l2p9h\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.490977 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pl4l\" (UniqueName: \"kubernetes.io/projected/878ce0cf-8408-4bd3-90cc-948a66892c07-kube-api-access-8pl4l\") pod \"control-plane-machine-set-operator-78cbb6b69f-mxb2z\" (UID: \"878ce0cf-8408-4bd3-90cc-948a66892c07\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.491038 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qvvg\" (UniqueName: \"kubernetes.io/projected/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-kube-api-access-5qvvg\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.491087 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpd88\" (UniqueName: \"kubernetes.io/projected/7112c6d3-009f-4a78-b6c3-5d61eaae0940-kube-api-access-qpd88\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.491147 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/25a09c7e-68bf-4a29-9bd5-909468bdcc42-ca-trust-extracted\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.491168 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j967\" (UniqueName: \"kubernetes.io/projected/df037075-04d9-4eab-bbc9-86ede7236798-kube-api-access-7j967\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.491194 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-etcd-ca\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.491293 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7112c6d3-009f-4a78-b6c3-5d61eaae0940-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: E1208 19:24:24.491334 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:24.991319974 +0000 UTC m=+147.633520977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.491361 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-certificates\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.492022 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs5b8\" (UniqueName: \"kubernetes.io/projected/98db57c6-9269-4d83-abd2-d71c34a20f23-kube-api-access-fs5b8\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.492059 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df037075-04d9-4eab-bbc9-86ede7236798-serving-cert\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.492091 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n2bsv\" (UID: \"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.524190 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" event={"ID":"979bd1cb-70e1-4f2d-b623-aaae40466dff","Type":"ContainerStarted","Data":"26ecfa47e41147d5b5e8d62fab925acb3569a9e751b3db51bb748332c180de07"} Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.531978 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" event={"ID":"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf","Type":"ContainerStarted","Data":"5cdcbb16d30e26197192ac37b3826639b4bdb8841290c8d2c84a3df8ccc5752c"} Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.553436 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.553638 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.555933 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.556934 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" event={"ID":"aca2089e-f16a-4bf6-9b7b-f3261a525cc3","Type":"ContainerStarted","Data":"cc8baff7719e0cd7304c7a97e817a895dbfecb67f9f51672e83dda819d8b7553"} Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.601753 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm"] Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.605680 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.605917 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-config\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.605950 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bed16b63-2523-4c5e-88b9-390cac136d1b-profile-collector-cert\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606010 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-bound-sa-token\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606030 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/47ae1194-5292-4938-8df3-cfcd4638b18e-proxy-tls\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606055 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-tls\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606070 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-trusted-ca\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606089 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7fb8e6fd-b0cb-42d5-9a59-6126db761239-signing-key\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606108 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606150 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/47ae1194-5292-4938-8df3-cfcd4638b18e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606189 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd519428-ba61-4a7b-98dd-45be1d9f1a30-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606205 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/47ae1194-5292-4938-8df3-cfcd4638b18e-images\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606222 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/98db57c6-9269-4d83-abd2-d71c34a20f23-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9n54\" (UniqueName: \"kubernetes.io/projected/bed16b63-2523-4c5e-88b9-390cac136d1b-kube-api-access-q9n54\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606355 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppz7x\" (UniqueName: \"kubernetes.io/projected/8693ec42-f6ae-4cf8-bcde-52843820ecd5-kube-api-access-ppz7x\") pod \"migrator-59844c95c7-gz9d5\" (UID: \"8693ec42-f6ae-4cf8-bcde-52843820ecd5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606371 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606400 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82da14d7-a213-4d6c-85f9-3f6f62e48323-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606417 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-stats-auth\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606444 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/25a09c7e-68bf-4a29-9bd5-909468bdcc42-installation-pull-secrets\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df037075-04d9-4eab-bbc9-86ede7236798-etcd-client\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606489 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/98db57c6-9269-4d83-abd2-d71c34a20f23-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606514 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7112c6d3-009f-4a78-b6c3-5d61eaae0940-srv-cert\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606529 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82da14d7-a213-4d6c-85f9-3f6f62e48323-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606548 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-registration-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606563 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-default-certificate\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606583 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2gz5\" (UniqueName: \"kubernetes.io/projected/d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6-kube-api-access-s2gz5\") pod \"package-server-manager-789f6589d5-n2bsv\" (UID: \"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606600 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7qdp\" (UniqueName: \"kubernetes.io/projected/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-kube-api-access-t7qdp\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606626 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqlfx\" (UniqueName: \"kubernetes.io/projected/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-kube-api-access-dqlfx\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606653 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhj48\" (UniqueName: \"kubernetes.io/projected/47ae1194-5292-4938-8df3-cfcd4638b18e-kube-api-access-hhj48\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606696 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bed16b63-2523-4c5e-88b9-390cac136d1b-srv-cert\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: E1208 19:24:24.606764 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.106733048 +0000 UTC m=+147.748934051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606865 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-node-bootstrap-token\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606915 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-metrics-certs\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606961 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-etcd-service-ca\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606979 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98db57c6-9269-4d83-abd2-d71c34a20f23-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.606999 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607019 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnt4z\" (UniqueName: \"kubernetes.io/projected/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-kube-api-access-lnt4z\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607041 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607061 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82da14d7-a213-4d6c-85f9-3f6f62e48323-config\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607079 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607098 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/878ce0cf-8408-4bd3-90cc-948a66892c07-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mxb2z\" (UID: \"878ce0cf-8408-4bd3-90cc-948a66892c07\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607135 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2p9h\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-kube-api-access-l2p9h\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607151 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-metrics-tls\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607190 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pl4l\" (UniqueName: \"kubernetes.io/projected/878ce0cf-8408-4bd3-90cc-948a66892c07-kube-api-access-8pl4l\") pod \"control-plane-machine-set-operator-78cbb6b69f-mxb2z\" (UID: \"878ce0cf-8408-4bd3-90cc-948a66892c07\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607229 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-config-volume\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607246 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7fb8e6fd-b0cb-42d5-9a59-6126db761239-signing-cabundle\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607321 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qvvg\" (UniqueName: \"kubernetes.io/projected/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-kube-api-access-5qvvg\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607369 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpd88\" (UniqueName: \"kubernetes.io/projected/7112c6d3-009f-4a78-b6c3-5d61eaae0940-kube-api-access-qpd88\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607386 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-socket-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607402 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-csi-data-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607524 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/25a09c7e-68bf-4a29-9bd5-909468bdcc42-ca-trust-extracted\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607543 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j967\" (UniqueName: \"kubernetes.io/projected/df037075-04d9-4eab-bbc9-86ede7236798-kube-api-access-7j967\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607563 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8nck\" (UniqueName: \"kubernetes.io/projected/97de0170-91b1-4c09-a75b-96ea32dfe71e-kube-api-access-l8nck\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607613 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6khb7\" (UniqueName: \"kubernetes.io/projected/014f906c-b438-42f0-8ea7-64342bbe492b-kube-api-access-6khb7\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607647 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-etcd-ca\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607669 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/483e2da1-f77e-49d1-9cf8-38fc2d3810ee-cert\") pod \"ingress-canary-f7kfm\" (UID: \"483e2da1-f77e-49d1-9cf8-38fc2d3810ee\") " pod="openshift-ingress-canary/ingress-canary-f7kfm" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607754 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/014f906c-b438-42f0-8ea7-64342bbe492b-machine-approver-tls\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607784 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607801 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/014f906c-b438-42f0-8ea7-64342bbe492b-config\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607842 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7112c6d3-009f-4a78-b6c3-5d61eaae0940-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607870 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-mountpoint-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607926 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/014f906c-b438-42f0-8ea7-64342bbe492b-auth-proxy-config\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607948 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-certificates\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607967 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs5b8\" (UniqueName: \"kubernetes.io/projected/98db57c6-9269-4d83-abd2-d71c34a20f23-kube-api-access-fs5b8\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.607984 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-service-ca-bundle\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608003 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd519428-ba61-4a7b-98dd-45be1d9f1a30-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608019 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-certs\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608046 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df037075-04d9-4eab-bbc9-86ede7236798-serving-cert\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608064 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n2bsv\" (UID: \"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608082 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-plugins-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608101 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x284j\" (UniqueName: \"kubernetes.io/projected/483e2da1-f77e-49d1-9cf8-38fc2d3810ee-kube-api-access-x284j\") pod \"ingress-canary-f7kfm\" (UID: \"483e2da1-f77e-49d1-9cf8-38fc2d3810ee\") " pod="openshift-ingress-canary/ingress-canary-f7kfm" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608154 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmwcx\" (UniqueName: \"kubernetes.io/projected/7fb8e6fd-b0cb-42d5-9a59-6126db761239-kube-api-access-kmwcx\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608183 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp4bf\" (UniqueName: \"kubernetes.io/projected/fd519428-ba61-4a7b-98dd-45be1d9f1a30-kube-api-access-rp4bf\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.608845 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-config\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.621615 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.630035 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/25a09c7e-68bf-4a29-9bd5-909468bdcc42-installation-pull-secrets\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.634970 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/25a09c7e-68bf-4a29-9bd5-909468bdcc42-ca-trust-extracted\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.636574 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-trusted-ca\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.639391 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/98db57c6-9269-4d83-abd2-d71c34a20f23-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.640104 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-certificates\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.640695 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-etcd-ca\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.650385 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/df037075-04d9-4eab-bbc9-86ede7236798-etcd-service-ca\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.651780 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.651928 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98db57c6-9269-4d83-abd2-d71c34a20f23-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.652934 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82da14d7-a213-4d6c-85f9-3f6f62e48323-config\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.657726 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.659304 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/878ce0cf-8408-4bd3-90cc-948a66892c07-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mxb2z\" (UID: \"878ce0cf-8408-4bd3-90cc-948a66892c07\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.660058 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df037075-04d9-4eab-bbc9-86ede7236798-etcd-client\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.660567 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n2bsv\" (UID: \"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.662811 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7112c6d3-009f-4a78-b6c3-5d61eaae0940-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.662845 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.663174 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7112c6d3-009f-4a78-b6c3-5d61eaae0940-srv-cert\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.663384 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bed16b63-2523-4c5e-88b9-390cac136d1b-profile-collector-cert\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.663814 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df037075-04d9-4eab-bbc9-86ede7236798-serving-cert\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.664029 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.664479 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bed16b63-2523-4c5e-88b9-390cac136d1b-srv-cert\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.664645 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-tls\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.665955 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9n54\" (UniqueName: \"kubernetes.io/projected/bed16b63-2523-4c5e-88b9-390cac136d1b-kube-api-access-q9n54\") pod \"catalog-operator-68c6474976-l7tjv\" (UID: \"bed16b63-2523-4c5e-88b9-390cac136d1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.667922 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.673576 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82da14d7-a213-4d6c-85f9-3f6f62e48323-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.680256 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.683978 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.684851 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppz7x\" (UniqueName: \"kubernetes.io/projected/8693ec42-f6ae-4cf8-bcde-52843820ecd5-kube-api-access-ppz7x\") pod \"migrator-59844c95c7-gz9d5\" (UID: \"8693ec42-f6ae-4cf8-bcde-52843820ecd5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.686499 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/98db57c6-9269-4d83-abd2-d71c34a20f23-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.705797 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82da14d7-a213-4d6c-85f9-3f6f62e48323-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-whpwl\" (UID: \"82da14d7-a213-4d6c-85f9-3f6f62e48323\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.709974 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-registration-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710009 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-default-certificate\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710035 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqlfx\" (UniqueName: \"kubernetes.io/projected/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-kube-api-access-dqlfx\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710114 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7qdp\" (UniqueName: \"kubernetes.io/projected/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-kube-api-access-t7qdp\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710136 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhj48\" (UniqueName: \"kubernetes.io/projected/47ae1194-5292-4938-8df3-cfcd4638b18e-kube-api-access-hhj48\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710159 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-node-bootstrap-token\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710183 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-metrics-certs\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710203 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnt4z\" (UniqueName: \"kubernetes.io/projected/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-kube-api-access-lnt4z\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710229 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-metrics-tls\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710296 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-config-volume\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710330 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7fb8e6fd-b0cb-42d5-9a59-6126db761239-signing-cabundle\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710373 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-socket-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710391 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-csi-data-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710430 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8nck\" (UniqueName: \"kubernetes.io/projected/97de0170-91b1-4c09-a75b-96ea32dfe71e-kube-api-access-l8nck\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710451 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6khb7\" (UniqueName: \"kubernetes.io/projected/014f906c-b438-42f0-8ea7-64342bbe492b-kube-api-access-6khb7\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710469 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/483e2da1-f77e-49d1-9cf8-38fc2d3810ee-cert\") pod \"ingress-canary-f7kfm\" (UID: \"483e2da1-f77e-49d1-9cf8-38fc2d3810ee\") " pod="openshift-ingress-canary/ingress-canary-f7kfm" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710491 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/014f906c-b438-42f0-8ea7-64342bbe492b-machine-approver-tls\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710528 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/014f906c-b438-42f0-8ea7-64342bbe492b-config\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710549 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-mountpoint-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710570 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/014f906c-b438-42f0-8ea7-64342bbe492b-auth-proxy-config\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710599 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd519428-ba61-4a7b-98dd-45be1d9f1a30-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710623 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-certs\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710663 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-service-ca-bundle\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710689 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-plugins-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710714 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x284j\" (UniqueName: \"kubernetes.io/projected/483e2da1-f77e-49d1-9cf8-38fc2d3810ee-kube-api-access-x284j\") pod \"ingress-canary-f7kfm\" (UID: \"483e2da1-f77e-49d1-9cf8-38fc2d3810ee\") " pod="openshift-ingress-canary/ingress-canary-f7kfm" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710740 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmwcx\" (UniqueName: \"kubernetes.io/projected/7fb8e6fd-b0cb-42d5-9a59-6126db761239-kube-api-access-kmwcx\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710767 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp4bf\" (UniqueName: \"kubernetes.io/projected/fd519428-ba61-4a7b-98dd-45be1d9f1a30-kube-api-access-rp4bf\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710815 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/47ae1194-5292-4938-8df3-cfcd4638b18e-proxy-tls\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710850 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710874 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7fb8e6fd-b0cb-42d5-9a59-6126db761239-signing-key\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710902 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/47ae1194-5292-4938-8df3-cfcd4638b18e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710928 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd519428-ba61-4a7b-98dd-45be1d9f1a30-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710953 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/47ae1194-5292-4938-8df3-cfcd4638b18e-images\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.710982 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-stats-auth\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.711446 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-registration-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.712514 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/014f906c-b438-42f0-8ea7-64342bbe492b-config\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.713371 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-mountpoint-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.715144 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-config-volume\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.715285 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd519428-ba61-4a7b-98dd-45be1d9f1a30-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.716075 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-plugins-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.717100 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7fb8e6fd-b0cb-42d5-9a59-6126db761239-signing-cabundle\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.717171 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-socket-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.717239 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/97de0170-91b1-4c09-a75b-96ea32dfe71e-csi-data-dir\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.718819 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-default-certificate\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.722659 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-service-ca-bundle\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.721444 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/014f906c-b438-42f0-8ea7-64342bbe492b-auth-proxy-config\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.723279 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/47ae1194-5292-4938-8df3-cfcd4638b18e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: E1208 19:24:24.723892 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.223849945 +0000 UTC m=+147.866051008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.726136 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/47ae1194-5292-4938-8df3-cfcd4638b18e-images\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.730228 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-stats-auth\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.731431 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.744816 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-metrics-certs\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.747708 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/483e2da1-f77e-49d1-9cf8-38fc2d3810ee-cert\") pod \"ingress-canary-f7kfm\" (UID: \"483e2da1-f77e-49d1-9cf8-38fc2d3810ee\") " pod="openshift-ingress-canary/ingress-canary-f7kfm" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.749214 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-node-bootstrap-token\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.749675 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/47ae1194-5292-4938-8df3-cfcd4638b18e-proxy-tls\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.751397 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/014f906c-b438-42f0-8ea7-64342bbe492b-machine-approver-tls\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.752020 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd519428-ba61-4a7b-98dd-45be1d9f1a30-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.753626 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-metrics-tls\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.753890 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpd88\" (UniqueName: \"kubernetes.io/projected/7112c6d3-009f-4a78-b6c3-5d61eaae0940-kube-api-access-qpd88\") pod \"olm-operator-6b444d44fb-2bghh\" (UID: \"7112c6d3-009f-4a78-b6c3-5d61eaae0940\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.754436 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7fb8e6fd-b0cb-42d5-9a59-6126db761239-signing-key\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.754937 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-certs\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.762162 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j967\" (UniqueName: \"kubernetes.io/projected/df037075-04d9-4eab-bbc9-86ede7236798-kube-api-access-7j967\") pod \"etcd-operator-b45778765-8d2q2\" (UID: \"df037075-04d9-4eab-bbc9-86ede7236798\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.773898 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2gz5\" (UniqueName: \"kubernetes.io/projected/d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6-kube-api-access-s2gz5\") pod \"package-server-manager-789f6589d5-n2bsv\" (UID: \"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.795245 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs5b8\" (UniqueName: \"kubernetes.io/projected/98db57c6-9269-4d83-abd2-d71c34a20f23-kube-api-access-fs5b8\") pod \"cluster-image-registry-operator-dc59b4c8b-fpkzq\" (UID: \"98db57c6-9269-4d83-abd2-d71c34a20f23\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.810338 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pl4l\" (UniqueName: \"kubernetes.io/projected/878ce0cf-8408-4bd3-90cc-948a66892c07-kube-api-access-8pl4l\") pod \"control-plane-machine-set-operator-78cbb6b69f-mxb2z\" (UID: \"878ce0cf-8408-4bd3-90cc-948a66892c07\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.813814 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:24 crc kubenswrapper[4706]: E1208 19:24:24.814063 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.314014016 +0000 UTC m=+147.956215019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.814558 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: E1208 19:24:24.815111 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.315091929 +0000 UTC m=+147.957292932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.826144 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qvvg\" (UniqueName: \"kubernetes.io/projected/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-kube-api-access-5qvvg\") pod \"marketplace-operator-79b997595-55zvr\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.835006 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.845488 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-bound-sa-token\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.853633 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.876255 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2p9h\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-kube-api-access-l2p9h\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.884663 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.905884 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.912214 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqlfx\" (UniqueName: \"kubernetes.io/projected/bf1c86d7-123d-472e-9fe0-a7644dd46ee2-kube-api-access-dqlfx\") pod \"machine-config-server-8v5f4\" (UID: \"bf1c86d7-123d-472e-9fe0-a7644dd46ee2\") " pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.916397 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:24 crc kubenswrapper[4706]: E1208 19:24:24.917019 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.41699912 +0000 UTC m=+148.059200123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.927070 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.938809 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmwcx\" (UniqueName: \"kubernetes.io/projected/7fb8e6fd-b0cb-42d5-9a59-6126db761239-kube-api-access-kmwcx\") pod \"service-ca-9c57cc56f-rzswq\" (UID: \"7fb8e6fd-b0cb-42d5-9a59-6126db761239\") " pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.939318 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.947530 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.965141 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8nck\" (UniqueName: \"kubernetes.io/projected/97de0170-91b1-4c09-a75b-96ea32dfe71e-kube-api-access-l8nck\") pod \"csi-hostpathplugin-q9lfk\" (UID: \"97de0170-91b1-4c09-a75b-96ea32dfe71e\") " pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.987024 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6khb7\" (UniqueName: \"kubernetes.io/projected/014f906c-b438-42f0-8ea7-64342bbe492b-kube-api-access-6khb7\") pod \"machine-approver-56656f9798-dpkz7\" (UID: \"014f906c-b438-42f0-8ea7-64342bbe492b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:24 crc kubenswrapper[4706]: I1208 19:24:24.996537 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.005097 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.011886 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnt4z\" (UniqueName: \"kubernetes.io/projected/d6be5b57-e25c-4ea6-9900-bbd5b72f2604-kube-api-access-lnt4z\") pod \"dns-default-ktnbl\" (UID: \"d6be5b57-e25c-4ea6-9900-bbd5b72f2604\") " pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.012824 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.022909 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.023340 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.523325936 +0000 UTC m=+148.165526939 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.038746 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.039739 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x284j\" (UniqueName: \"kubernetes.io/projected/483e2da1-f77e-49d1-9cf8-38fc2d3810ee-kube-api-access-x284j\") pod \"ingress-canary-f7kfm\" (UID: \"483e2da1-f77e-49d1-9cf8-38fc2d3810ee\") " pod="openshift-ingress-canary/ingress-canary-f7kfm" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.045309 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp4bf\" (UniqueName: \"kubernetes.io/projected/fd519428-ba61-4a7b-98dd-45be1d9f1a30-kube-api-access-rp4bf\") pod \"kube-storage-version-migrator-operator-b67b599dd-zgrj8\" (UID: \"fd519428-ba61-4a7b-98dd-45be1d9f1a30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.049682 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.055478 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7qdp\" (UniqueName: \"kubernetes.io/projected/65a505d5-7286-4b1c-ad40-f2e83c1e39d2-kube-api-access-t7qdp\") pod \"router-default-5444994796-cvvxf\" (UID: \"65a505d5-7286-4b1c-ad40-f2e83c1e39d2\") " pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.057761 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-f7kfm" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.074800 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8v5f4" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.074963 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhj48\" (UniqueName: \"kubernetes.io/projected/47ae1194-5292-4938-8df3-cfcd4638b18e-kube-api-access-hhj48\") pod \"machine-config-operator-74547568cd-fqkdp\" (UID: \"47ae1194-5292-4938-8df3-cfcd4638b18e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.077655 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.086630 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.089539 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gs7dv"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.098481 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.122447 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.124735 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.125122 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl"] Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.125766 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.62565676 +0000 UTC m=+148.267857763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.138201 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.138888 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.638869295 +0000 UTC m=+148.281070298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.140619 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.181311 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.192904 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zgngm"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.239935 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.245873 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.74583602 +0000 UTC m=+148.388037023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.260663 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bxc4t"] Dec 08 19:24:25 crc kubenswrapper[4706]: W1208 19:24:25.268318 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4124e67a_ef7b_4984_bb29_29525802aa66.slice/crio-9791ac009006062cb76e0e841911b054af2bc451c0cd93e2d4aa3ef97193ff5d WatchSource:0}: Error finding container 9791ac009006062cb76e0e841911b054af2bc451c0cd93e2d4aa3ef97193ff5d: Status 404 returned error can't find the container with id 9791ac009006062cb76e0e841911b054af2bc451c0cd93e2d4aa3ef97193ff5d Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.277039 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6rqvg"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.281560 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gv9fs"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.320097 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.327653 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.344932 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.346699 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.347097 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.847073661 +0000 UTC m=+148.489274664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.451677 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.452063 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:25.952010254 +0000 UTC m=+148.594211257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.554168 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.554984 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.054969127 +0000 UTC m=+148.697170130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.657660 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.658030 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.158002572 +0000 UTC m=+148.800203575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.658324 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.660816 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.160795658 +0000 UTC m=+148.802996661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: W1208 19:24:25.673474 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98220d47_1a06_4e17_90ec_f802f041197c.slice/crio-bb5f1620b9e5f660c57c429c960fadaf5c1c3349cf92deb189808b821b9d2a89 WatchSource:0}: Error finding container bb5f1620b9e5f660c57c429c960fadaf5c1c3349cf92deb189808b821b9d2a89: Status 404 returned error can't find the container with id bb5f1620b9e5f660c57c429c960fadaf5c1c3349cf92deb189808b821b9d2a89 Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.697702 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" event={"ID":"979bd1cb-70e1-4f2d-b623-aaae40466dff","Type":"ContainerStarted","Data":"1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.697811 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.698823 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" event={"ID":"aca2089e-f16a-4bf6-9b7b-f3261a525cc3","Type":"ContainerStarted","Data":"a50d15f7684537fb7d2559aa3cfae9a5de73833a0b0c6d6576b224d1d58a8715"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.699790 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-gs7dv" event={"ID":"0f5724d1-fcce-4453-9ca0-8d38414cae8f","Type":"ContainerStarted","Data":"1b59cb5b10d6824984382ab041426e664a7756bf1afe67c6603ee5944c512566"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.700530 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" event={"ID":"014f906c-b438-42f0-8ea7-64342bbe492b","Type":"ContainerStarted","Data":"85e44eddb05d2693ee8cb1e6b048087d9982bbc091c6500488f1635ea9d61975"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.705454 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.706560 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" event={"ID":"37e229d7-a86c-4e5b-bf51-caec4edc7c39","Type":"ContainerStarted","Data":"4eeb8fcab0a6e1ac63ca72728deafb80fc24d84645ff260a2107a1c8024aec85"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.726741 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" event={"ID":"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf","Type":"ContainerStarted","Data":"4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.726842 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.729186 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" event={"ID":"0e775a27-59a7-4c73-a2ae-7598d29b91fc","Type":"ContainerStarted","Data":"713356f2bc0f4df1b030e2a0a6a1f4026bfe12e040ab85f0e2c16c5dc99ac83f"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.730204 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" event={"ID":"4124e67a-ef7b-4984-bb29-29525802aa66","Type":"ContainerStarted","Data":"9791ac009006062cb76e0e841911b054af2bc451c0cd93e2d4aa3ef97193ff5d"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.731331 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" event={"ID":"e0fd938d-859f-4620-944c-598cbe045a1a","Type":"ContainerStarted","Data":"8ba848a964cc72206c1e199f0385835d0799d29cc4205c80433e7bfc0e26327e"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.732201 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8v5f4" event={"ID":"bf1c86d7-123d-472e-9fe0-a7644dd46ee2","Type":"ContainerStarted","Data":"fc2d532bf7b86c6732cedc5815eb071c5dc34c350fb19c403f85eafb0000f2f0"} Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.768335 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.768839 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.268819446 +0000 UTC m=+148.911020449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.795852 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.803653 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.820678 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.872701 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.877534 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.377513945 +0000 UTC m=+149.019714948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.949456 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr"] Dec 08 19:24:25 crc kubenswrapper[4706]: I1208 19:24:25.973735 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:25 crc kubenswrapper[4706]: E1208 19:24:25.974189 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.474163435 +0000 UTC m=+149.116364438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.085789 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.086321 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.586299539 +0000 UTC m=+149.228500542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.186463 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.190238 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.690214531 +0000 UTC m=+149.332415534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.292908 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.293301 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.793285977 +0000 UTC m=+149.435486980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: W1208 19:24:26.374466 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod035c9071_5a5b_4204_810d_1ee403423d01.slice/crio-707af17042d8053c17270b135557df8b06d4c184bd336c3a3cf333af2e686a75 WatchSource:0}: Error finding container 707af17042d8053c17270b135557df8b06d4c184bd336c3a3cf333af2e686a75: Status 404 returned error can't find the container with id 707af17042d8053c17270b135557df8b06d4c184bd336c3a3cf333af2e686a75 Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.396332 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.396783 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:26.896761405 +0000 UTC m=+149.538962408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.458000 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.489877 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n"] Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.509396 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.509837 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.009822028 +0000 UTC m=+149.652023031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.588118 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" podStartSLOduration=125.588089945 podStartE2EDuration="2m5.588089945s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:26.545394077 +0000 UTC m=+149.187595080" watchObservedRunningTime="2025-12-08 19:24:26.588089945 +0000 UTC m=+149.230290948" Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.591205 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" podStartSLOduration=126.591183159 podStartE2EDuration="2m6.591183159s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:26.583855025 +0000 UTC m=+149.226056028" watchObservedRunningTime="2025-12-08 19:24:26.591183159 +0000 UTC m=+149.233384162" Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.614820 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.615437 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.115412471 +0000 UTC m=+149.757613474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.725684 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.726153 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.226135842 +0000 UTC m=+149.868336845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: W1208 19:24:26.748637 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe5d2ba7_9cfa_4952_a142_e15839642dd9.slice/crio-007bf993b12013ce8806958f4432b616f7a9d3b459c8971449d402461d98a961 WatchSource:0}: Error finding container 007bf993b12013ce8806958f4432b616f7a9d3b459c8971449d402461d98a961: Status 404 returned error can't find the container with id 007bf993b12013ce8806958f4432b616f7a9d3b459c8971449d402461d98a961 Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.777709 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" event={"ID":"aca2089e-f16a-4bf6-9b7b-f3261a525cc3","Type":"ContainerStarted","Data":"74a8e5376d15ea14457db16eb54e4127a17b08cc1e635d1f891863d195072d96"} Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.810883 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" event={"ID":"c0d44c3b-0f39-4db6-9553-cf23c89fe209","Type":"ContainerStarted","Data":"15ba462a537bc6925d592da146e843fda8c21e4c585cc272802c3e95b8063ed1"} Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.811710 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vh2jf" podStartSLOduration=125.811687102 podStartE2EDuration="2m5.811687102s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:26.810793975 +0000 UTC m=+149.452994978" watchObservedRunningTime="2025-12-08 19:24:26.811687102 +0000 UTC m=+149.453888105" Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.829241 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.829793 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.329773926 +0000 UTC m=+149.971974929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.838634 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bxc4t" event={"ID":"7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb","Type":"ContainerStarted","Data":"aa130f822af7163009f68567816511d2212b4448631441199b9ad5c9a9a25692"} Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.881827 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" event={"ID":"69a6c982-b070-4f4a-8b0c-f841fc147a50","Type":"ContainerStarted","Data":"0ec243a5c0b5208a2c067a0421a5ac9d3bf8b3b99b1fa4d1862063a7fb9f0144"} Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.906014 4706 generic.go:334] "Generic (PLEG): container finished" podID="e0fd938d-859f-4620-944c-598cbe045a1a" containerID="4634cc779b47c2bdc8653ac05f35c73918a449fc7369fda406f8effa1600ec99" exitCode=0 Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.922431 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" event={"ID":"e0fd938d-859f-4620-944c-598cbe045a1a","Type":"ContainerDied","Data":"4634cc779b47c2bdc8653ac05f35c73918a449fc7369fda406f8effa1600ec99"} Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.933916 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" event={"ID":"eaed88d5-42bb-478a-842e-1326e77fbc1e","Type":"ContainerStarted","Data":"68fc73f8ff7dfa7c03000dbc03aa50737849b2b392ac99f2fe309c092110aa6b"} Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.938735 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-gs7dv" event={"ID":"0f5724d1-fcce-4453-9ca0-8d38414cae8f","Type":"ContainerStarted","Data":"092c325d6797cddfd06627a2e7bfd3b8882cc4100675aa3744255c4cdbec9ea5"} Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.939822 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.948726 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:26 crc kubenswrapper[4706]: E1208 19:24:26.949550 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.449525013 +0000 UTC m=+150.091726016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:26 crc kubenswrapper[4706]: I1208 19:24:26.974457 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:26.999698 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.000587 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-gs7dv" podStartSLOduration=127.000566666 podStartE2EDuration="2m7.000566666s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:26.974772406 +0000 UTC m=+149.616973409" watchObservedRunningTime="2025-12-08 19:24:27.000566666 +0000 UTC m=+149.642767679" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.011344 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j6s78"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.033850 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" event={"ID":"37e229d7-a86c-4e5b-bf51-caec4edc7c39","Type":"ContainerStarted","Data":"331e5fcba2c3c5a40b451bfff88d8e40782582ddc2882b1fdfd85799cfb45084"} Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.072200 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.073333 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.573118568 +0000 UTC m=+150.215319571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.124044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" event={"ID":"98220d47-1a06-4e17-90ec-f802f041197c","Type":"ContainerStarted","Data":"bb5f1620b9e5f660c57c429c960fadaf5c1c3349cf92deb189808b821b9d2a89"} Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.126043 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.126375 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" podStartSLOduration=127.126354528 podStartE2EDuration="2m7.126354528s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:27.124745819 +0000 UTC m=+149.766946822" watchObservedRunningTime="2025-12-08 19:24:27.126354528 +0000 UTC m=+149.768555531" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.144826 4706 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-pk89q container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.144907 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" podUID="98220d47-1a06-4e17-90ec-f802f041197c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.159974 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.168555 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.181302 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zgngm" event={"ID":"3f1e0eea-50be-4dd8-a504-dcdd049ae255","Type":"ContainerStarted","Data":"9c19b71f75c4ea9c54aa8af463b5b88662d8075b5ebbb11681b338e52223667b"} Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.183292 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" podStartSLOduration=126.183245331 podStartE2EDuration="2m6.183245331s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:27.182207229 +0000 UTC m=+149.824408232" watchObservedRunningTime="2025-12-08 19:24:27.183245331 +0000 UTC m=+149.825446334" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.183649 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.187979 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.687956515 +0000 UTC m=+150.330157518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.219936 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rzswq"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.228538 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8d2q2"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.229709 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" event={"ID":"db465e9d-aadc-41bc-ba53-3fb564d4a541","Type":"ContainerStarted","Data":"3bedc8b18d1419340714f3c5cd93b058d0b60ce43503a4a5b4fc31c70f09f91c"} Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.264498 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" event={"ID":"6889021e-b838-4fb3-b664-92efcdf699a2","Type":"ContainerStarted","Data":"3a21e665f812116df409a65c5eb5ea469187996a5872dc2169d67a52769f5dfd"} Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.280969 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" podStartSLOduration=126.280952283 podStartE2EDuration="2m6.280952283s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:27.278865289 +0000 UTC m=+149.921066282" watchObservedRunningTime="2025-12-08 19:24:27.280952283 +0000 UTC m=+149.923153286" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.283434 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" event={"ID":"035c9071-5a5b-4204-810d-1ee403423d01","Type":"ContainerStarted","Data":"707af17042d8053c17270b135557df8b06d4c184bd336c3a3cf333af2e686a75"} Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.293468 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.296624 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.796590572 +0000 UTC m=+150.438791575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.313727 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2mptf"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.336033 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" podStartSLOduration=127.335997798 podStartE2EDuration="2m7.335997798s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:27.327377305 +0000 UTC m=+149.969578308" watchObservedRunningTime="2025-12-08 19:24:27.335997798 +0000 UTC m=+149.978198801" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.377152 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" event={"ID":"ef62dd12-f35d-4808-a54b-da8fb928592a","Type":"ContainerStarted","Data":"8f8090566af02604ad6bbbceef51c11d89f2ae2a615d273eb5b587115ba6eb86"} Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.382498 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8v5f4" podStartSLOduration=6.382477142 podStartE2EDuration="6.382477142s" podCreationTimestamp="2025-12-08 19:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:27.379064797 +0000 UTC m=+150.021265800" watchObservedRunningTime="2025-12-08 19:24:27.382477142 +0000 UTC m=+150.024678155" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.396684 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.397523 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cvvxf" event={"ID":"65a505d5-7286-4b1c-ad40-f2e83c1e39d2","Type":"ContainerStarted","Data":"c2590ee3d749e926af27449c37d090f876e190026c8f6de245dd23a185505a73"} Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.402729 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.902707131 +0000 UTC m=+150.544908134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.497883 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.499079 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:27.999057572 +0000 UTC m=+150.641258575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.504410 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-cvvxf" podStartSLOduration=126.504387655 podStartE2EDuration="2m6.504387655s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:27.431870195 +0000 UTC m=+150.074071198" watchObservedRunningTime="2025-12-08 19:24:27.504387655 +0000 UTC m=+150.146588658" Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.504655 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.550854 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-f7kfm"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.613014 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.613632 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.1136126 +0000 UTC m=+150.755813603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.696471 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.696977 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-55zvr"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.715034 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.715546 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.215524251 +0000 UTC m=+150.857725254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.715728 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ktnbl"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.744393 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.762004 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.772153 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.786973 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.793413 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-q9lfk"] Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.801089 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv"] Dec 08 19:24:27 crc kubenswrapper[4706]: W1208 19:24:27.801853 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b144ce7_6f82_46a9_ac9d_0a98aa48bbac.slice/crio-daf051574b2b7e7c71e14d25cf0108c1c857d59903d3623649f23ca7b725975e WatchSource:0}: Error finding container daf051574b2b7e7c71e14d25cf0108c1c857d59903d3623649f23ca7b725975e: Status 404 returned error can't find the container with id daf051574b2b7e7c71e14d25cf0108c1c857d59903d3623649f23ca7b725975e Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.822921 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.823420 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.323404585 +0000 UTC m=+150.965605588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.828596 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-gs7dv" Dec 08 19:24:27 crc kubenswrapper[4706]: W1208 19:24:27.864439 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6be5b57_e25c_4ea6_9900_bbd5b72f2604.slice/crio-048f721342659916d0cfdc4820fe24101832aa44180409e097c4c1d9019fc199 WatchSource:0}: Error finding container 048f721342659916d0cfdc4820fe24101832aa44180409e097c4c1d9019fc199: Status 404 returned error can't find the container with id 048f721342659916d0cfdc4820fe24101832aa44180409e097c4c1d9019fc199 Dec 08 19:24:27 crc kubenswrapper[4706]: W1208 19:24:27.915156 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd519428_ba61_4a7b_98dd_45be1d9f1a30.slice/crio-69295e0bab9cde1a04ea30602e1202a9f1ef9f09003f6fa56b2d8a8795476c2c WatchSource:0}: Error finding container 69295e0bab9cde1a04ea30602e1202a9f1ef9f09003f6fa56b2d8a8795476c2c: Status 404 returned error can't find the container with id 69295e0bab9cde1a04ea30602e1202a9f1ef9f09003f6fa56b2d8a8795476c2c Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.924977 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.925503 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.42546683 +0000 UTC m=+151.067667833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: I1208 19:24:27.925684 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:27 crc kubenswrapper[4706]: E1208 19:24:27.926044 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.426030568 +0000 UTC m=+151.068231561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:27 crc kubenswrapper[4706]: W1208 19:24:27.974775 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3b6d176_7fe1_4f7b_8fa8_612cfd68cdb6.slice/crio-221c73346c553b3db9494ce0644c4c964cb93109e8b7e1c4cf554053f6299540 WatchSource:0}: Error finding container 221c73346c553b3db9494ce0644c4c964cb93109e8b7e1c4cf554053f6299540: Status 404 returned error can't find the container with id 221c73346c553b3db9494ce0644c4c964cb93109e8b7e1c4cf554053f6299540 Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.033828 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.034944 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.534912732 +0000 UTC m=+151.177113735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.136643 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.137016 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.637002119 +0000 UTC m=+151.279203122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.237336 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.237701 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.737684012 +0000 UTC m=+151.379885015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.339284 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.339912 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.839799469 +0000 UTC m=+151.482000472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.348557 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.364785 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:28 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:28 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:28 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.364864 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.440619 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.441200 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:28.941163523 +0000 UTC m=+151.583364526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.441818 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cvvxf" event={"ID":"65a505d5-7286-4b1c-ad40-f2e83c1e39d2","Type":"ContainerStarted","Data":"f88d2de818f0b45a319f5811b1adb3fef6baf32e6e3f84409447c61fd61a6085"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.467605 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" event={"ID":"82da14d7-a213-4d6c-85f9-3f6f62e48323","Type":"ContainerStarted","Data":"20fe62c27b5fd665247e23cfbc0bac46d475c2ebfd1d13af3a96a66e8d4dc3cf"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.469769 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" event={"ID":"df037075-04d9-4eab-bbc9-86ede7236798","Type":"ContainerStarted","Data":"28f780e4ccb92ed6289f35fa9efbba73bea57fdfe7a60f3ed4677c61f9a4f349"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.540212 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" event={"ID":"bed16b63-2523-4c5e-88b9-390cac136d1b","Type":"ContainerStarted","Data":"7c1af0451dbecf948cf073fa7fc2dee59ffefa85b3bebddc7ca10e3274817b17"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.540741 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" event={"ID":"bed16b63-2523-4c5e-88b9-390cac136d1b","Type":"ContainerStarted","Data":"fa7c030f850ec81cbdf8354c3b9fcea90cab0a96df8fee2ce4c2e9bc36528e04"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.540763 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.543060 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.543528 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.043515938 +0000 UTC m=+151.685716941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.548192 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bxc4t" event={"ID":"7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb","Type":"ContainerStarted","Data":"2489a428bd213f3bc29770804222dfb8fa3d539dc1459b3dd34cfe513906a30f"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.548290 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bxc4t" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.554856 4706 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-l7tjv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.554910 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" podUID="bed16b63-2523-4c5e-88b9-390cac136d1b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.555117 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" event={"ID":"8693ec42-f6ae-4cf8-bcde-52843820ecd5","Type":"ContainerStarted","Data":"242031c278755390abd8193e03a2e15fb5826860a8588aea2f6e30503c03ad1e"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.555142 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" event={"ID":"8693ec42-f6ae-4cf8-bcde-52843820ecd5","Type":"ContainerStarted","Data":"9172707a03d1c88ff6fe43599fc9e90ef5e30651a3b5e7b78d5be788dca29baa"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.558813 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ktnbl" event={"ID":"d6be5b57-e25c-4ea6-9900-bbd5b72f2604","Type":"ContainerStarted","Data":"048f721342659916d0cfdc4820fe24101832aa44180409e097c4c1d9019fc199"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.563579 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-bxc4t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.563643 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bxc4t" podUID="7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.573711 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-f7kfm" event={"ID":"483e2da1-f77e-49d1-9cf8-38fc2d3810ee","Type":"ContainerStarted","Data":"5b09b0fd17f2002e598b2ea974b6416aa1422b01ac0f62e96b1bcd38403fbd93"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.614575 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" event={"ID":"0e775a27-59a7-4c73-a2ae-7598d29b91fc","Type":"ContainerStarted","Data":"e028c38c718ded41eeec28739fe6d8510adbcbbcbac615677fec2bcd272e4a1d"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.638209 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" podStartSLOduration=127.638184947 podStartE2EDuration="2m7.638184947s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:28.634900726 +0000 UTC m=+151.277101729" watchObservedRunningTime="2025-12-08 19:24:28.638184947 +0000 UTC m=+151.280385950" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.644708 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.645350 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.145321815 +0000 UTC m=+151.787522818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.646098 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.664874 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.164853293 +0000 UTC m=+151.807054296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.676679 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" event={"ID":"ef62dd12-f35d-4808-a54b-da8fb928592a","Type":"ContainerStarted","Data":"2bc78e4460cfdbe57d4fec45f6b2a43d4a10637d65193ddfca231c1a5d99116b"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.676752 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" event={"ID":"ef62dd12-f35d-4808-a54b-da8fb928592a","Type":"ContainerStarted","Data":"cbd414e61bd3a75869169e350d3e95dbb4dcf0b28e92c6f40ff90ca91a7be677"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.703597 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bxc4t" podStartSLOduration=127.703575559 podStartE2EDuration="2m7.703575559s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:28.668752303 +0000 UTC m=+151.310953306" watchObservedRunningTime="2025-12-08 19:24:28.703575559 +0000 UTC m=+151.345776562" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.739078 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"51255ca82fa03a158d12d1c3b0e7c73f7b9434335aac75d512f0c709215d3a05"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.746931 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:28 crc kubenswrapper[4706]: E1208 19:24:28.748552 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.248498225 +0000 UTC m=+151.890699238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.762502 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fddbl" podStartSLOduration=127.762477343 podStartE2EDuration="2m7.762477343s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:28.708472339 +0000 UTC m=+151.350673342" watchObservedRunningTime="2025-12-08 19:24:28.762477343 +0000 UTC m=+151.404678346" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.769037 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ggc6b" podStartSLOduration=127.769021763 podStartE2EDuration="2m7.769021763s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:28.758509312 +0000 UTC m=+151.400710315" watchObservedRunningTime="2025-12-08 19:24:28.769021763 +0000 UTC m=+151.411222976" Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.770189 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" event={"ID":"eaed88d5-42bb-478a-842e-1326e77fbc1e","Type":"ContainerStarted","Data":"e45c3475e27842e2aff8723c1db0eaba99004f084841e27d02a3102d45952bd9"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.770236 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" event={"ID":"eaed88d5-42bb-478a-842e-1326e77fbc1e","Type":"ContainerStarted","Data":"4f4c8fe5ec6f4a8c55b8d700a27f587f164ed1b5614f93df498ce48958e39c6b"} Dec 08 19:24:28 crc kubenswrapper[4706]: I1208 19:24:28.840112 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" event={"ID":"7112c6d3-009f-4a78-b6c3-5d61eaae0940","Type":"ContainerStarted","Data":"1372f9144af54fa8c778cfa6f95b4a5b6aa8279f79509c15eef48a33939a82c9"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.848867 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:28.866748 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.366726166 +0000 UTC m=+152.008927169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.870409 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-gv9fs" podStartSLOduration=127.870392448 podStartE2EDuration="2m7.870392448s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:28.866829669 +0000 UTC m=+151.509030672" watchObservedRunningTime="2025-12-08 19:24:28.870392448 +0000 UTC m=+151.512593451" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.877141 4706 generic.go:334] "Generic (PLEG): container finished" podID="4124e67a-ef7b-4984-bb29-29525802aa66" containerID="334cbf79ed7177e62ac934b3678bea848abab47f6eaf9357de323162d08130c5" exitCode=0 Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.880341 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" event={"ID":"4124e67a-ef7b-4984-bb29-29525802aa66","Type":"ContainerDied","Data":"334cbf79ed7177e62ac934b3678bea848abab47f6eaf9357de323162d08130c5"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.910905 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"240b161ebbadb3ab85a9eb79056fa0b889e53959d5d3ad81d7cfd8eeeba57009"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.910966 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ceed83596c35a4a501468ec9c86098f3674d4ae78118a895b56aae9475a8c392"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.914734 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5b6e7ecc98c7a050fbbba5ff56f37bb6801185c957207f55c3af4ef8da40328b"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.919952 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" event={"ID":"035c9071-5a5b-4204-810d-1ee403423d01","Type":"ContainerStarted","Data":"f6abf3988d7feab6c381ad06c48b4cd76cc488a00cfba4e14f830ae54b5a3931"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.949988 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:28.950357 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.450335686 +0000 UTC m=+152.092536689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:28.990170 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" event={"ID":"97de0170-91b1-4c09-a75b-96ea32dfe71e","Type":"ContainerStarted","Data":"a61aded901a4f5f89fa8ed2957f2d08591bea753baaa0c93407891f802025b4f"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.017389 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" event={"ID":"878ce0cf-8408-4bd3-90cc-948a66892c07","Type":"ContainerStarted","Data":"fc3daf62b03965115bd704bdce3330363a2bdf9705b22d9433f4e9966f1bb117"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.052400 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.055973 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.555954981 +0000 UTC m=+152.198155984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.103611 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" event={"ID":"014f906c-b438-42f0-8ea7-64342bbe492b","Type":"ContainerStarted","Data":"ebda02ccf994b1489c116c894c4a3003d3b22a3ac412737a8596525a22cdcf98"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.122979 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" event={"ID":"98220d47-1a06-4e17-90ec-f802f041197c","Type":"ContainerStarted","Data":"dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.136813 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-8h7dr" podStartSLOduration=128.136787706 podStartE2EDuration="2m8.136787706s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.008339412 +0000 UTC m=+151.650540425" watchObservedRunningTime="2025-12-08 19:24:29.136787706 +0000 UTC m=+151.778988719" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.136860 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.156131 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.156731 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.656697626 +0000 UTC m=+152.298898629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.165510 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" event={"ID":"7fb8e6fd-b0cb-42d5-9a59-6126db761239","Type":"ContainerStarted","Data":"1e950030ad7f15d70c656e390c485ba26f71d4fd682ca91ce8b9d76927da9b56"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.165581 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" event={"ID":"7fb8e6fd-b0cb-42d5-9a59-6126db761239","Type":"ContainerStarted","Data":"d0c890cbd6c837ae26eeeee34ea81c2e0d91853f82ff893e123fe52ed8c9d434"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.170740 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" podStartSLOduration=129.170712845 podStartE2EDuration="2m9.170712845s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.137322422 +0000 UTC m=+151.779523425" watchObservedRunningTime="2025-12-08 19:24:29.170712845 +0000 UTC m=+151.812913848" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.179411 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" event={"ID":"4ae483b4-35fe-4790-9523-c5e6b1fd3137","Type":"ContainerStarted","Data":"4387039716faad29e0504b884d5755c7a987e592396cbf6c94105555464c1077"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.179485 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" event={"ID":"4ae483b4-35fe-4790-9523-c5e6b1fd3137","Type":"ContainerStarted","Data":"60ff33c04a11e8c56fca2aa7c913ab4412cebcdc2c4c775b6586f92976171259"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.219751 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rzswq" podStartSLOduration=128.219729466 podStartE2EDuration="2m8.219729466s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.21791518 +0000 UTC m=+151.860116173" watchObservedRunningTime="2025-12-08 19:24:29.219729466 +0000 UTC m=+151.861930459" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.229718 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgtpm" event={"ID":"37e229d7-a86c-4e5b-bf51-caec4edc7c39","Type":"ContainerStarted","Data":"350ebf6b00c52fec6ada1154a848eeefaacf1ee17ced505b745e0c5b3c97ce98"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.247170 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" event={"ID":"98db57c6-9269-4d83-abd2-d71c34a20f23","Type":"ContainerStarted","Data":"fe9eda2fd55216baefa46ce2a1d1cb517fe0ab117e855923541f5b8136c6ea7c"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.259332 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.264187 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.764166967 +0000 UTC m=+152.406367970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.269673 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qr4hz" podStartSLOduration=129.269654695 podStartE2EDuration="2m9.269654695s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.267716976 +0000 UTC m=+151.909917979" watchObservedRunningTime="2025-12-08 19:24:29.269654695 +0000 UTC m=+151.911855698" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.278416 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zgngm" event={"ID":"3f1e0eea-50be-4dd8-a504-dcdd049ae255","Type":"ContainerStarted","Data":"3da593dd8d025eec1bab4fbb886c67739cf46dba49a7488da0684e80e6385c29"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.321869 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" podStartSLOduration=128.321827723 podStartE2EDuration="2m8.321827723s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.31488169 +0000 UTC m=+151.957082693" watchObservedRunningTime="2025-12-08 19:24:29.321827723 +0000 UTC m=+151.964028726" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.323612 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" event={"ID":"69a6c982-b070-4f4a-8b0c-f841fc147a50","Type":"ContainerStarted","Data":"1d616617ea20ffd758a7ab5891fe7a43366c68454985b4ddaa1c383e51211040"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.326813 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.349179 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" event={"ID":"e0fd938d-859f-4620-944c-598cbe045a1a","Type":"ContainerStarted","Data":"05ccf309f12ea0568e25a2a81cb0e0d7fa7536a1cd710fe7062d9119c44307a5"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.349567 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.356589 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.358689 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-zgngm" podStartSLOduration=128.358660371 podStartE2EDuration="2m8.358660371s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.356755732 +0000 UTC m=+151.998956735" watchObservedRunningTime="2025-12-08 19:24:29.358660371 +0000 UTC m=+152.000861374" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.360587 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.362169 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.862145267 +0000 UTC m=+152.504346280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.362507 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:29 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:29 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:29 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.362609 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.399098 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" event={"ID":"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac","Type":"ContainerStarted","Data":"daf051574b2b7e7c71e14d25cf0108c1c857d59903d3623649f23ca7b725975e"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.433957 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9nzm8" podStartSLOduration=128.433937716 podStartE2EDuration="2m8.433937716s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.432214123 +0000 UTC m=+152.074415136" watchObservedRunningTime="2025-12-08 19:24:29.433937716 +0000 UTC m=+152.076138719" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.442970 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" event={"ID":"323ba66c-aab0-4d31-982b-b012b399bc5c","Type":"ContainerStarted","Data":"d7eebe32614c7a507706b5faa74e3d03ddc11f875679a7e89b40c0704a4b62af"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.443033 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" event={"ID":"323ba66c-aab0-4d31-982b-b012b399bc5c","Type":"ContainerStarted","Data":"aba98242b32a3059b3227199e2ba3113650affac5c002aa46f1b268e96f7d7c3"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.466076 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.468351 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:29.968330729 +0000 UTC m=+152.610531812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.484712 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" event={"ID":"47ae1194-5292-4938-8df3-cfcd4638b18e","Type":"ContainerStarted","Data":"ff94b9d7311af287782e509643663a2a0379a4ec7439b0e0f30830154b6d48b9"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.501713 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8v5f4" event={"ID":"bf1c86d7-123d-472e-9fe0-a7644dd46ee2","Type":"ContainerStarted","Data":"e2de981d9e973d84f3da1b25c6ee32513f8b15e89590d08e8a12d9373de0b2c8"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.518815 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" podStartSLOduration=129.518774664 podStartE2EDuration="2m9.518774664s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.517928368 +0000 UTC m=+152.160129361" watchObservedRunningTime="2025-12-08 19:24:29.518774664 +0000 UTC m=+152.160975667" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.528144 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" event={"ID":"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6","Type":"ContainerStarted","Data":"221c73346c553b3db9494ce0644c4c964cb93109e8b7e1c4cf554053f6299540"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.561511 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" event={"ID":"be5d2ba7-9cfa-4952-a142-e15839642dd9","Type":"ContainerStarted","Data":"b5513c17436fe22b34f1bd3fb4581e9ac84fdab6e7808ba961ddcc9c7af820a9"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.561561 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" event={"ID":"be5d2ba7-9cfa-4952-a142-e15839642dd9","Type":"ContainerStarted","Data":"007bf993b12013ce8806958f4432b616f7a9d3b459c8971449d402461d98a961"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.568773 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.570329 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.070257771 +0000 UTC m=+152.712458774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.611641 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" event={"ID":"6889021e-b838-4fb3-b664-92efcdf699a2","Type":"ContainerStarted","Data":"3c30807b9e78027ef9f14262817bcc633e951ecc33add8e00cf7d96213ed582f"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.678049 4706 generic.go:334] "Generic (PLEG): container finished" podID="c0d44c3b-0f39-4db6-9553-cf23c89fe209" containerID="f3034cc9da72c0af8fa60ecb0c02c2e951343a8653485fd2ec86c26e37540fdd" exitCode=0 Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.685116 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.685674 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.185654605 +0000 UTC m=+152.827855608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.707440 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sg66n" podStartSLOduration=128.707418191 podStartE2EDuration="2m8.707418191s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.590523581 +0000 UTC m=+152.232724604" watchObservedRunningTime="2025-12-08 19:24:29.707418191 +0000 UTC m=+152.349619194" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.707718 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" podStartSLOduration=128.7077135 podStartE2EDuration="2m8.7077135s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.704637156 +0000 UTC m=+152.346838179" watchObservedRunningTime="2025-12-08 19:24:29.7077135 +0000 UTC m=+152.349914503" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.778662 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" event={"ID":"fd519428-ba61-4a7b-98dd-45be1d9f1a30","Type":"ContainerStarted","Data":"69295e0bab9cde1a04ea30602e1202a9f1ef9f09003f6fa56b2d8a8795476c2c"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.778727 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" event={"ID":"48960e60-2160-4afa-98ca-e67603fc9545","Type":"ContainerStarted","Data":"eba76cce546d16aa5b714e6f8150cac38c01e21c80c9f023713f7c71d72df1b6"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.778747 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" event={"ID":"48960e60-2160-4afa-98ca-e67603fc9545","Type":"ContainerStarted","Data":"2580839a9d48b2e75b4bb94aa1a2e71adfcfe9495b5a43a176977a1884eb7033"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.778760 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2rns4" event={"ID":"db465e9d-aadc-41bc-ba53-3fb564d4a541","Type":"ContainerStarted","Data":"ea9607b92e2800de4824a49fbbd9c9b153264a7c493a7570b21de737dad684d7"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.778797 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" event={"ID":"c0d44c3b-0f39-4db6-9553-cf23c89fe209","Type":"ContainerDied","Data":"f3034cc9da72c0af8fa60ecb0c02c2e951343a8653485fd2ec86c26e37540fdd"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.778813 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" event={"ID":"dacda7e8-5813-40a1-bf74-b09eb69c4a40","Type":"ContainerStarted","Data":"235cd4e356b9aad19c70006f4ac3a39811801db5e0a6f43e2c773dec93db2e2c"} Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.782009 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" podStartSLOduration=129.781984624 podStartE2EDuration="2m9.781984624s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:29.774736273 +0000 UTC m=+152.416937276" watchObservedRunningTime="2025-12-08 19:24:29.781984624 +0000 UTC m=+152.424185627" Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.786473 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.786560 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.286539964 +0000 UTC m=+152.928740967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.787099 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.796520 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.296501939 +0000 UTC m=+152.938702942 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.889983 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.890429 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.390409435 +0000 UTC m=+153.032610438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:29 crc kubenswrapper[4706]: I1208 19:24:29.991460 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:29 crc kubenswrapper[4706]: E1208 19:24:29.991818 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.491805219 +0000 UTC m=+153.134006212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.092390 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.093545 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.593522864 +0000 UTC m=+153.235723867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.139124 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tkq2v" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.195173 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.195583 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.695569469 +0000 UTC m=+153.337770472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.296155 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.296482 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.796440138 +0000 UTC m=+153.438641151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.298502 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.299450 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.79942131 +0000 UTC m=+153.441622313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.368315 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:30 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:30 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:30 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.368772 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.400168 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.400609 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:30.900586258 +0000 UTC m=+153.542787261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.502852 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.503447 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.003427287 +0000 UTC m=+153.645628290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.604977 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.605962 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.105928396 +0000 UTC m=+153.748129399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.622117 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2h6vv"] Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.626806 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.630555 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.651959 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2h6vv"] Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.708371 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-catalog-content\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.708450 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxkgj\" (UniqueName: \"kubernetes.io/projected/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-kube-api-access-mxkgj\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.708569 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-utilities\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.708638 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.709129 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.209111656 +0000 UTC m=+153.851312659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.779452 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6s78" event={"ID":"48960e60-2160-4afa-98ca-e67603fc9545","Type":"ContainerStarted","Data":"5e8c91a5c541c3de5e369b7b0d3eafa7ab20e40cad1c4a46f5b3e6e45c75db61"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.796518 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ktnbl" event={"ID":"d6be5b57-e25c-4ea6-9900-bbd5b72f2604","Type":"ContainerStarted","Data":"388f3210b49e9d5d0bc2b08adc00ca2572f1f8d8d72945e5988cd83cc70cc056"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.808040 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q6t9t"] Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.809251 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.809426 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.809781 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-utilities\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.809839 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-catalog-content\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.809860 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxkgj\" (UniqueName: \"kubernetes.io/projected/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-kube-api-access-mxkgj\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.810722 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-utilities\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.811661 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-catalog-content\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: W1208 19:24:30.811913 4706 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.811949 4706 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.812609 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.312569874 +0000 UTC m=+153.954770877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.812633 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"edf803edd2bcc52361da7b6985de80b4efb6d15e4f345c6294dbb21e204d7138"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.826881 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fpkzq" event={"ID":"98db57c6-9269-4d83-abd2-d71c34a20f23","Type":"ContainerStarted","Data":"f74fd314bdb47bc31f5d78cdefd4f4cebdde2ec9cecef6aa7a4103264a1991d8"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.837875 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q6t9t"] Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.859990 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" event={"ID":"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6","Type":"ContainerStarted","Data":"e8912996f77e4fa8fc604c922d8c4d12df1067f0645e1b3fa7a047add015179e"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.870431 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxkgj\" (UniqueName: \"kubernetes.io/projected/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-kube-api-access-mxkgj\") pod \"certified-operators-2h6vv\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.872870 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"62d7f95a7c96d5a348412c29eda4c9393de2196729dad35f23f3626352921ab3"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.873615 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.897101 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" event={"ID":"c0d44c3b-0f39-4db6-9553-cf23c89fe209","Type":"ContainerStarted","Data":"052de3a368a37be96f79bd1def1e51fca9113d647562af4b45d432b070081b27"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.913499 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" event={"ID":"7112c6d3-009f-4a78-b6c3-5d61eaae0940","Type":"ContainerStarted","Data":"2a346e6f0759a815f207a2b96a70bae27e219dded55dba7181e6dca66ba079ce"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.914253 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-utilities\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.914349 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.914370 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w7m4\" (UniqueName: \"kubernetes.io/projected/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-kube-api-access-5w7m4\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.914431 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-catalog-content\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.914780 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:30 crc kubenswrapper[4706]: E1208 19:24:30.915777 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.415760714 +0000 UTC m=+154.057961707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.936230 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" event={"ID":"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac","Type":"ContainerStarted","Data":"9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.937478 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.937826 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.947003 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" event={"ID":"fd519428-ba61-4a7b-98dd-45be1d9f1a30","Type":"ContainerStarted","Data":"93d3df0dd2e301e5be092f4c355097b3e928fda95f4008b63b45793cfa9c902a"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.950158 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-55zvr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.950242 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" podUID="2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.959380 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-f7kfm" event={"ID":"483e2da1-f77e-49d1-9cf8-38fc2d3810ee","Type":"ContainerStarted","Data":"16a1bf3a81cf57f8168a4fd257a2e9df9991e2e2f207f716b4c5555782bb56a5"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.964829 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2bghh" podStartSLOduration=129.964801116 podStartE2EDuration="2m9.964801116s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:30.962112814 +0000 UTC m=+153.604313817" watchObservedRunningTime="2025-12-08 19:24:30.964801116 +0000 UTC m=+153.607002119" Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.973243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" event={"ID":"47ae1194-5292-4938-8df3-cfcd4638b18e","Type":"ContainerStarted","Data":"1f1aaaa7cf4fb0982da0648f32f5aea7da43ca0c36b3abcb2e7657879f7727af"} Dec 08 19:24:30 crc kubenswrapper[4706]: I1208 19:24:30.992454 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" event={"ID":"df037075-04d9-4eab-bbc9-86ede7236798","Type":"ContainerStarted","Data":"3ba05d19714797c33a30d6c7dc9146d33feccab8ece5ba21d9160ff5b134299d"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.019660 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-njntc"] Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.020874 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.024040 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.024446 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-catalog-content\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.024571 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-utilities\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.024724 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w7m4\" (UniqueName: \"kubernetes.io/projected/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-kube-api-access-5w7m4\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.026841 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-utilities\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.027892 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.527849247 +0000 UTC m=+154.170050260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.030257 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-catalog-content\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.037496 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zgrj8" podStartSLOduration=130.037479252 podStartE2EDuration="2m10.037479252s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.036862463 +0000 UTC m=+153.679063466" watchObservedRunningTime="2025-12-08 19:24:31.037479252 +0000 UTC m=+153.679680255" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.041609 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dpkz7" event={"ID":"014f906c-b438-42f0-8ea7-64342bbe492b","Type":"ContainerStarted","Data":"6b2c7a79782a5d325f620bab7edbeb8ad7be411d171161025022a03e2c017bfc"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.053396 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w7m4\" (UniqueName: \"kubernetes.io/projected/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-kube-api-access-5w7m4\") pod \"community-operators-q6t9t\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.053925 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.055044 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-njntc"] Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.073316 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" event={"ID":"323ba66c-aab0-4d31-982b-b012b399bc5c","Type":"ContainerStarted","Data":"f514026db8ab0c96aedfbf0d7769a1c286bb64a18acc7c17b33468c5f6d8320d"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.075807 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" podStartSLOduration=130.075769064 podStartE2EDuration="2m10.075769064s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.058828026 +0000 UTC m=+153.701029029" watchObservedRunningTime="2025-12-08 19:24:31.075769064 +0000 UTC m=+153.717970077" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.135915 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rgbr\" (UniqueName: \"kubernetes.io/projected/0a9bae31-9b10-44fb-b30f-7cfca35cda15-kube-api-access-4rgbr\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.136634 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-catalog-content\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.136704 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-utilities\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.136963 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.137620 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.637608168 +0000 UTC m=+154.279809161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.145819 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" event={"ID":"878ce0cf-8408-4bd3-90cc-948a66892c07","Type":"ContainerStarted","Data":"b71a1f2ff15e5529798b0e6c4a8cb4d6b55134e72ec76c64ec49299bf1c94fe1"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.179520 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-f7kfm" podStartSLOduration=10.179485531 podStartE2EDuration="10.179485531s" podCreationTimestamp="2025-12-08 19:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.114025856 +0000 UTC m=+153.756226859" watchObservedRunningTime="2025-12-08 19:24:31.179485531 +0000 UTC m=+153.821686534" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.189331 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2mptf" event={"ID":"dacda7e8-5813-40a1-bf74-b09eb69c4a40","Type":"ContainerStarted","Data":"427df917157e47e11b2da794d4703fffcad807dfdd5f6f6cd4d6fbd341e143db"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.214890 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" podStartSLOduration=130.214865564 podStartE2EDuration="2m10.214865564s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.172795286 +0000 UTC m=+153.814996309" watchObservedRunningTime="2025-12-08 19:24:31.214865564 +0000 UTC m=+153.857066567" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.218349 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-97l9j" podStartSLOduration=130.21833687 podStartE2EDuration="2m10.21833687s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.204785065 +0000 UTC m=+153.846986068" watchObservedRunningTime="2025-12-08 19:24:31.21833687 +0000 UTC m=+153.860537873" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.222673 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8rl9w"] Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.233033 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" event={"ID":"4124e67a-ef7b-4984-bb29-29525802aa66","Type":"ContainerStarted","Data":"49c7d1849693b8afc708cfe760231f1748a1eff8f1362d2af5449f1c269290c9"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.233168 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.233291 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" event={"ID":"82da14d7-a213-4d6c-85f9-3f6f62e48323","Type":"ContainerStarted","Data":"3964c694c8f489d2bde6f52af9b92f971b96dfd81faabf478f446d1e170fb51d"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.238761 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.239144 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rgbr\" (UniqueName: \"kubernetes.io/projected/0a9bae31-9b10-44fb-b30f-7cfca35cda15-kube-api-access-4rgbr\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.239177 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-catalog-content\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.239207 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-utilities\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.240473 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.740451898 +0000 UTC m=+154.382652901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.241053 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-catalog-content\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.241281 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-utilities\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.252455 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8rl9w"] Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.261828 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8d2q2" podStartSLOduration=130.261807332 podStartE2EDuration="2m10.261807332s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.261498602 +0000 UTC m=+153.903699605" watchObservedRunningTime="2025-12-08 19:24:31.261807332 +0000 UTC m=+153.904008325" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.309750 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" event={"ID":"8693ec42-f6ae-4cf8-bcde-52843820ecd5","Type":"ContainerStarted","Data":"53b6abc3dd0215092c8a30914b19e25e8c338d7bd9060b85460fd2caf670b350"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.314490 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rgbr\" (UniqueName: \"kubernetes.io/projected/0a9bae31-9b10-44fb-b30f-7cfca35cda15-kube-api-access-4rgbr\") pod \"certified-operators-njntc\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.353480 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-whpwl" podStartSLOduration=130.353458488 podStartE2EDuration="2m10.353458488s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.351961203 +0000 UTC m=+153.994162206" watchObservedRunningTime="2025-12-08 19:24:31.353458488 +0000 UTC m=+153.995659481" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.354678 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-catalog-content\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.354728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.354761 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-utilities\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.354833 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5pxf\" (UniqueName: \"kubernetes.io/projected/3200b927-522a-4762-a731-c56780f4a137-kube-api-access-d5pxf\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.366531 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" event={"ID":"97de0170-91b1-4c09-a75b-96ea32dfe71e","Type":"ContainerStarted","Data":"2a7148694d13ba99044c57c56ad10f8f5ce2043acc6b189a2f17950516e9aa57"} Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.375954 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:31 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:31 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:31 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.376011 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.377436 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.877417012 +0000 UTC m=+154.519618015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.431975 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-bxc4t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.432083 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bxc4t" podUID="7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.455309 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.464075 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.465861 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-catalog-content\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.466014 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-utilities\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.466066 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5pxf\" (UniqueName: \"kubernetes.io/projected/3200b927-522a-4762-a731-c56780f4a137-kube-api-access-d5pxf\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.467819 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-catalog-content\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.469515 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-utilities\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.470174 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7tjv" Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.475066 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:31.9661825 +0000 UTC m=+154.608383503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.478542 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mxb2z" podStartSLOduration=130.478512178 podStartE2EDuration="2m10.478512178s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.43317923 +0000 UTC m=+154.075380233" watchObservedRunningTime="2025-12-08 19:24:31.478512178 +0000 UTC m=+154.120713181" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.540694 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" podStartSLOduration=130.540673042 podStartE2EDuration="2m10.540673042s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.53831924 +0000 UTC m=+154.180520253" watchObservedRunningTime="2025-12-08 19:24:31.540673042 +0000 UTC m=+154.182874045" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.551732 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5pxf\" (UniqueName: \"kubernetes.io/projected/3200b927-522a-4762-a731-c56780f4a137-kube-api-access-d5pxf\") pod \"community-operators-8rl9w\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.587362 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.601579 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.101548626 +0000 UTC m=+154.743749629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.611074 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gz9d5" podStartSLOduration=130.611045247 podStartE2EDuration="2m10.611045247s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:31.596692717 +0000 UTC m=+154.238893720" watchObservedRunningTime="2025-12-08 19:24:31.611045247 +0000 UTC m=+154.253246240" Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.689848 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.690412 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.190392177 +0000 UTC m=+154.832593180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.705293 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2h6vv"] Dec 08 19:24:31 crc kubenswrapper[4706]: W1208 19:24:31.777127 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod443b0c73_a6b1_4f48_b46c_7eae5ff8c7c5.slice/crio-7a13e5d4c360407b6435bc3d7ff3ef8b0f82d645711859ae9a55f1e839674c62 WatchSource:0}: Error finding container 7a13e5d4c360407b6435bc3d7ff3ef8b0f82d645711859ae9a55f1e839674c62: Status 404 returned error can't find the container with id 7a13e5d4c360407b6435bc3d7ff3ef8b0f82d645711859ae9a55f1e839674c62 Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.796401 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.798849 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.298831688 +0000 UTC m=+154.941032691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:31 crc kubenswrapper[4706]: I1208 19:24:31.904683 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:31 crc kubenswrapper[4706]: E1208 19:24:31.905074 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.405054791 +0000 UTC m=+155.047255794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.006201 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:32 crc kubenswrapper[4706]: E1208 19:24:32.006643 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.506628311 +0000 UTC m=+155.148829334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.033843 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-njntc"] Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.107621 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:32 crc kubenswrapper[4706]: E1208 19:24:32.108191 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.608170711 +0000 UTC m=+155.250371714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.161235 4706 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/community-operators-q6t9t" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.161365 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.210193 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:32 crc kubenswrapper[4706]: E1208 19:24:32.210887 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.710864826 +0000 UTC m=+155.353065829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.244884 4706 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.287275 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.295229 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.311647 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:32 crc kubenswrapper[4706]: E1208 19:24:32.312678 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.812654763 +0000 UTC m=+155.454855756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.359120 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:32 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:32 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:32 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.359187 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.389306 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ktnbl" event={"ID":"d6be5b57-e25c-4ea6-9900-bbd5b72f2604","Type":"ContainerStarted","Data":"4f782dedda99327877cafc85f38f02c5c16035d1987677d898d37ce85623c2a9"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.390930 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.413564 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:32 crc kubenswrapper[4706]: E1208 19:24:32.414006 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:32.913991356 +0000 UTC m=+155.556192359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.414328 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqkdp" event={"ID":"47ae1194-5292-4938-8df3-cfcd4638b18e","Type":"ContainerStarted","Data":"3ae8ef722681ab2f7ef6bbea322295f62cde14072e72b36dc5053e0bf9e43785"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.421034 4706 generic.go:334] "Generic (PLEG): container finished" podID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerID="cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275" exitCode=0 Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.421172 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h6vv" event={"ID":"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5","Type":"ContainerDied","Data":"cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.421207 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h6vv" event={"ID":"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5","Type":"ContainerStarted","Data":"7a13e5d4c360407b6435bc3d7ff3ef8b0f82d645711859ae9a55f1e839674c62"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.438944 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.441157 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" event={"ID":"d3b6d176-7fe1-4f7b-8fa8-612cfd68cdb6","Type":"ContainerStarted","Data":"ae8d62c4f59e3828d85fc181bf1d010c1ab41911c18be651daf27e8da0304b2a"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.442013 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.455596 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-ktnbl" podStartSLOduration=11.45557656 podStartE2EDuration="11.45557656s" podCreationTimestamp="2025-12-08 19:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:32.416626647 +0000 UTC m=+155.058827650" watchObservedRunningTime="2025-12-08 19:24:32.45557656 +0000 UTC m=+155.097777563" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.464698 4706 generic.go:334] "Generic (PLEG): container finished" podID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerID="65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf" exitCode=0 Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.464940 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-njntc" event={"ID":"0a9bae31-9b10-44fb-b30f-7cfca35cda15","Type":"ContainerDied","Data":"65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.464973 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-njntc" event={"ID":"0a9bae31-9b10-44fb-b30f-7cfca35cda15","Type":"ContainerStarted","Data":"7316c7e08caa2ef92dfff7f512cc163f90b7a01f018653a8386a8dfcf3a1dd41"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.469157 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" event={"ID":"97de0170-91b1-4c09-a75b-96ea32dfe71e","Type":"ContainerStarted","Data":"16e8ce3c961bbaa1444f35c4902f72ff0f5f32d440ddbf7e33a206e201092a84"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.487388 4706 generic.go:334] "Generic (PLEG): container finished" podID="6889021e-b838-4fb3-b664-92efcdf699a2" containerID="3c30807b9e78027ef9f14262817bcc633e951ecc33add8e00cf7d96213ed582f" exitCode=0 Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.487461 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" event={"ID":"6889021e-b838-4fb3-b664-92efcdf699a2","Type":"ContainerDied","Data":"3c30807b9e78027ef9f14262817bcc633e951ecc33add8e00cf7d96213ed582f"} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.514113 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" podStartSLOduration=131.514090232 podStartE2EDuration="2m11.514090232s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:32.486245929 +0000 UTC m=+155.128446932" watchObservedRunningTime="2025-12-08 19:24:32.514090232 +0000 UTC m=+155.156291235" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.516903 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.516993 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-55zvr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.517041 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" podUID="2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.517504 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" event={"ID":"c0d44c3b-0f39-4db6-9553-cf23c89fe209","Type":"ContainerStarted","Data":"b8bc7a64855cd388684689672fd917d9efdbaa69337c7fc9eef59c579d6e210a"} Dec 08 19:24:32 crc kubenswrapper[4706]: E1208 19:24:32.518338 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:33.018299271 +0000 UTC m=+155.660500274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.578298 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" podStartSLOduration=132.578276597 podStartE2EDuration="2m12.578276597s" podCreationTimestamp="2025-12-08 19:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:32.575856553 +0000 UTC m=+155.218057556" watchObservedRunningTime="2025-12-08 19:24:32.578276597 +0000 UTC m=+155.220477600" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.621613 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:32 crc kubenswrapper[4706]: E1208 19:24:32.627631 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 19:24:33.127612388 +0000 UTC m=+155.769813391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-28lq2" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.670219 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q6t9t"] Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.723789 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:32 crc kubenswrapper[4706]: E1208 19:24:32.724664 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 19:24:33.22463964 +0000 UTC m=+155.866840643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.766945 4706 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-08T19:24:32.244918099Z","Handler":null,"Name":""} Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.788440 4706 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.788498 4706 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.803856 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6fmnd"] Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.805115 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.824637 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.826467 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.830032 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.830078 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.830252 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fmnd"] Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.898174 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8rl9w"] Dec 08 19:24:32 crc kubenswrapper[4706]: W1208 19:24:32.916728 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3200b927_522a_4762_a731_c56780f4a137.slice/crio-571f25d3aa36fb8ffe8d314844633b2404d868d9e9d3d89546dc56fcac181167 WatchSource:0}: Error finding container 571f25d3aa36fb8ffe8d314844633b2404d868d9e9d3d89546dc56fcac181167: Status 404 returned error can't find the container with id 571f25d3aa36fb8ffe8d314844633b2404d868d9e9d3d89546dc56fcac181167 Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.927316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-catalog-content\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.927357 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whp92\" (UniqueName: \"kubernetes.io/projected/26f226b0-fa6c-4efc-af99-f69c1ad8a464-kube-api-access-whp92\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.927393 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-utilities\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.930606 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-28lq2\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:32 crc kubenswrapper[4706]: I1208 19:24:32.979041 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.028526 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.029140 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-catalog-content\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.029171 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whp92\" (UniqueName: \"kubernetes.io/projected/26f226b0-fa6c-4efc-af99-f69c1ad8a464-kube-api-access-whp92\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.029206 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-utilities\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.029823 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-utilities\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.029984 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-catalog-content\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.043147 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.070866 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whp92\" (UniqueName: \"kubernetes.io/projected/26f226b0-fa6c-4efc-af99-f69c1ad8a464-kube-api-access-whp92\") pod \"redhat-marketplace-6fmnd\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.163720 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.206578 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zcrgl"] Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.207951 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.221255 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zcrgl"] Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.336038 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-utilities\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.336499 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-catalog-content\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.336541 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrvzb\" (UniqueName: \"kubernetes.io/projected/0415ede8-d911-4058-ba7f-507f756ef909-kube-api-access-jrvzb\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.365832 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:33 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:33 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:33 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.365899 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.439784 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-catalog-content\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.442380 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrvzb\" (UniqueName: \"kubernetes.io/projected/0415ede8-d911-4058-ba7f-507f756ef909-kube-api-access-jrvzb\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.442417 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-utilities\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.442318 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-catalog-content\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.443553 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-utilities\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.462609 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-28lq2"] Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.472377 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrvzb\" (UniqueName: \"kubernetes.io/projected/0415ede8-d911-4058-ba7f-507f756ef909-kube-api-access-jrvzb\") pod \"redhat-marketplace-zcrgl\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.552555 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.554521 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" event={"ID":"25a09c7e-68bf-4a29-9bd5-909468bdcc42","Type":"ContainerStarted","Data":"13266bccd0bad2d5a1c3694b87553b9ee81bd8213ce29b5123df49a387374ed3"} Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.556580 4706 generic.go:334] "Generic (PLEG): container finished" podID="3200b927-522a-4762-a731-c56780f4a137" containerID="a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b" exitCode=0 Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.556671 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rl9w" event={"ID":"3200b927-522a-4762-a731-c56780f4a137","Type":"ContainerDied","Data":"a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b"} Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.556709 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rl9w" event={"ID":"3200b927-522a-4762-a731-c56780f4a137","Type":"ContainerStarted","Data":"571f25d3aa36fb8ffe8d314844633b2404d868d9e9d3d89546dc56fcac181167"} Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.561631 4706 generic.go:334] "Generic (PLEG): container finished" podID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerID="1554d09d7a70534828b57a5c1ab61b370ecaf946344e71188cac78cff8e3403f" exitCode=0 Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.561711 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6t9t" event={"ID":"9752bb6b-117b-4e2b-8794-a91e14dbeaf7","Type":"ContainerDied","Data":"1554d09d7a70534828b57a5c1ab61b370ecaf946344e71188cac78cff8e3403f"} Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.561739 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6t9t" event={"ID":"9752bb6b-117b-4e2b-8794-a91e14dbeaf7","Type":"ContainerStarted","Data":"c6ff3ca1d15d49c553a618a7b2cc37e5a9787ba282073734b58cf041d88639f1"} Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.585943 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" event={"ID":"97de0170-91b1-4c09-a75b-96ea32dfe71e","Type":"ContainerStarted","Data":"7540a89d337372eefe5c45f17ccacae404eb22e2ecb6d16b07bfe6f958e10c7e"} Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.587229 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" event={"ID":"97de0170-91b1-4c09-a75b-96ea32dfe71e","Type":"ContainerStarted","Data":"baa1f56471d7e0bade39c58adbc53e0850a4cdc3e5127abaf3af1cad6e1bc370"} Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.597099 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fmnd"] Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.604382 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:24:33 crc kubenswrapper[4706]: W1208 19:24:33.642239 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26f226b0_fa6c_4efc_af99_f69c1ad8a464.slice/crio-11cfa80821e6de8c7cf71cce6010fa42d058c9925580148eb63baa5c95750f52 WatchSource:0}: Error finding container 11cfa80821e6de8c7cf71cce6010fa42d058c9925580148eb63baa5c95750f52: Status 404 returned error can't find the container with id 11cfa80821e6de8c7cf71cce6010fa42d058c9925580148eb63baa5c95750f52 Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.644537 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-q9lfk" podStartSLOduration=11.644517369 podStartE2EDuration="11.644517369s" podCreationTimestamp="2025-12-08 19:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:33.643702874 +0000 UTC m=+156.285903877" watchObservedRunningTime="2025-12-08 19:24:33.644517369 +0000 UTC m=+156.286718372" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.724101 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.812795 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zsdd9"] Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.814475 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.825410 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.857836 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zsdd9"] Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.971155 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-utilities\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.971215 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-catalog-content\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:33 crc kubenswrapper[4706]: I1208 19:24:33.971282 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxfrp\" (UniqueName: \"kubernetes.io/projected/63da40ba-a150-4396-bb8d-d8e4096a3961-kube-api-access-lxfrp\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.018365 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.073546 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-utilities\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.073633 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-catalog-content\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.073720 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxfrp\" (UniqueName: \"kubernetes.io/projected/63da40ba-a150-4396-bb8d-d8e4096a3961-kube-api-access-lxfrp\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.074862 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-catalog-content\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.076929 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-utilities\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.099336 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxfrp\" (UniqueName: \"kubernetes.io/projected/63da40ba-a150-4396-bb8d-d8e4096a3961-kube-api-access-lxfrp\") pod \"redhat-operators-zsdd9\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.107765 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zcrgl"] Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.143419 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.143549 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:34 crc kubenswrapper[4706]: W1208 19:24:34.151709 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0415ede8_d911_4058_ba7f_507f756ef909.slice/crio-1a4ba1c80c852d76fa79a53cbd27523737c25ea28c411e09e8206f17e36b9e20 WatchSource:0}: Error finding container 1a4ba1c80c852d76fa79a53cbd27523737c25ea28c411e09e8206f17e36b9e20: Status 404 returned error can't find the container with id 1a4ba1c80c852d76fa79a53cbd27523737c25ea28c411e09e8206f17e36b9e20 Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.155075 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.155114 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.159499 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.175594 4706 patch_prober.go:28] interesting pod/console-f9d7485db-zgngm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.175674 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zgngm" podUID="3f1e0eea-50be-4dd8-a504-dcdd049ae255" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.181010 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6889021e-b838-4fb3-b664-92efcdf699a2-config-volume\") pod \"6889021e-b838-4fb3-b664-92efcdf699a2\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.181881 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6889021e-b838-4fb3-b664-92efcdf699a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "6889021e-b838-4fb3-b664-92efcdf699a2" (UID: "6889021e-b838-4fb3-b664-92efcdf699a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.182018 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sj9p\" (UniqueName: \"kubernetes.io/projected/6889021e-b838-4fb3-b664-92efcdf699a2-kube-api-access-9sj9p\") pod \"6889021e-b838-4fb3-b664-92efcdf699a2\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.182800 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6889021e-b838-4fb3-b664-92efcdf699a2-secret-volume\") pod \"6889021e-b838-4fb3-b664-92efcdf699a2\" (UID: \"6889021e-b838-4fb3-b664-92efcdf699a2\") " Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.183667 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6889021e-b838-4fb3-b664-92efcdf699a2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.197657 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6889021e-b838-4fb3-b664-92efcdf699a2-kube-api-access-9sj9p" (OuterVolumeSpecName: "kube-api-access-9sj9p") pod "6889021e-b838-4fb3-b664-92efcdf699a2" (UID: "6889021e-b838-4fb3-b664-92efcdf699a2"). InnerVolumeSpecName "kube-api-access-9sj9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.207055 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.210465 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6889021e-b838-4fb3-b664-92efcdf699a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6889021e-b838-4fb3-b664-92efcdf699a2" (UID: "6889021e-b838-4fb3-b664-92efcdf699a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.213749 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-bxc4t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.213804 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bxc4t" podUID="7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.214177 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wwksc"] Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.214546 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-bxc4t container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.214573 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bxc4t" podUID="7dd08a83-51c4-4f21-93bc-d7e7b4d73dbb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 08 19:24:34 crc kubenswrapper[4706]: E1208 19:24:34.214798 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6889021e-b838-4fb3-b664-92efcdf699a2" containerName="collect-profiles" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.214818 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6889021e-b838-4fb3-b664-92efcdf699a2" containerName="collect-profiles" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.214922 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="6889021e-b838-4fb3-b664-92efcdf699a2" containerName="collect-profiles" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.215757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.215944 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwksc"] Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.221766 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.221830 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.241053 4706 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6rqvg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]log ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]etcd ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/generic-apiserver-start-informers ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/max-in-flight-filter ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 08 19:24:34 crc kubenswrapper[4706]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/project.openshift.io-projectcache ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/openshift.io-startinformers ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 08 19:24:34 crc kubenswrapper[4706]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 08 19:24:34 crc kubenswrapper[4706]: livez check failed Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.241152 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" podUID="c0d44c3b-0f39-4db6-9553-cf23c89fe209" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.285055 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sj9p\" (UniqueName: \"kubernetes.io/projected/6889021e-b838-4fb3-b664-92efcdf699a2-kube-api-access-9sj9p\") on node \"crc\" DevicePath \"\"" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.285105 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6889021e-b838-4fb3-b664-92efcdf699a2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.355990 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:34 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:34 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:34 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.356147 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.387076 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-utilities\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.387660 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsnb5\" (UniqueName: \"kubernetes.io/projected/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-kube-api-access-fsnb5\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.387969 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-catalog-content\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.493044 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-utilities\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.493156 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsnb5\" (UniqueName: \"kubernetes.io/projected/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-kube-api-access-fsnb5\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.493207 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-catalog-content\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.495131 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-catalog-content\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.495157 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-utilities\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.520646 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsnb5\" (UniqueName: \"kubernetes.io/projected/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-kube-api-access-fsnb5\") pod \"redhat-operators-wwksc\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.541007 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zsdd9"] Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.548592 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:24:34 crc kubenswrapper[4706]: W1208 19:24:34.553551 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63da40ba_a150_4396_bb8d_d8e4096a3961.slice/crio-58ebe40f9802ae8a6088b27eac961fb80c5cf52c35f1fb3be0384df5c8ea7f28 WatchSource:0}: Error finding container 58ebe40f9802ae8a6088b27eac961fb80c5cf52c35f1fb3be0384df5c8ea7f28: Status 404 returned error can't find the container with id 58ebe40f9802ae8a6088b27eac961fb80c5cf52c35f1fb3be0384df5c8ea7f28 Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.671473 4706 generic.go:334] "Generic (PLEG): container finished" podID="0415ede8-d911-4058-ba7f-507f756ef909" containerID="c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4" exitCode=0 Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.671584 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zcrgl" event={"ID":"0415ede8-d911-4058-ba7f-507f756ef909","Type":"ContainerDied","Data":"c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4"} Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.671622 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zcrgl" event={"ID":"0415ede8-d911-4058-ba7f-507f756ef909","Type":"ContainerStarted","Data":"1a4ba1c80c852d76fa79a53cbd27523737c25ea28c411e09e8206f17e36b9e20"} Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.676367 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" event={"ID":"6889021e-b838-4fb3-b664-92efcdf699a2","Type":"ContainerDied","Data":"3a21e665f812116df409a65c5eb5ea469187996a5872dc2169d67a52769f5dfd"} Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.676429 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a21e665f812116df409a65c5eb5ea469187996a5872dc2169d67a52769f5dfd" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.676375 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.681797 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zsdd9" event={"ID":"63da40ba-a150-4396-bb8d-d8e4096a3961","Type":"ContainerStarted","Data":"58ebe40f9802ae8a6088b27eac961fb80c5cf52c35f1fb3be0384df5c8ea7f28"} Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.686418 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" event={"ID":"25a09c7e-68bf-4a29-9bd5-909468bdcc42","Type":"ContainerStarted","Data":"422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990"} Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.686505 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.701039 4706 generic.go:334] "Generic (PLEG): container finished" podID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerID="2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac" exitCode=0 Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.702096 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fmnd" event={"ID":"26f226b0-fa6c-4efc-af99-f69c1ad8a464","Type":"ContainerDied","Data":"2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac"} Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.702127 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fmnd" event={"ID":"26f226b0-fa6c-4efc-af99-f69c1ad8a464","Type":"ContainerStarted","Data":"11cfa80821e6de8c7cf71cce6010fa42d058c9925580148eb63baa5c95750f52"} Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.712416 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" podStartSLOduration=133.712394982 podStartE2EDuration="2m13.712394982s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:34.706308636 +0000 UTC m=+157.348509639" watchObservedRunningTime="2025-12-08 19:24:34.712394982 +0000 UTC m=+157.354595985" Dec 08 19:24:34 crc kubenswrapper[4706]: I1208 19:24:34.718799 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t4qw2" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.012416 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.013940 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.019528 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.020889 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.023662 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.106420 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.106975 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.208399 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.208486 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.208592 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.229814 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.346827 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.351395 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.351665 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:35 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:35 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:35 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.351763 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.414013 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwksc"] Dec 08 19:24:35 crc kubenswrapper[4706]: W1208 19:24:35.461943 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32aaa66c_bc2d_42e3_9c17_681d58e0ef08.slice/crio-f3ef61765a27b583a52d7ccb7bed896f0887df696601635f046adce38e2d7ba7 WatchSource:0}: Error finding container f3ef61765a27b583a52d7ccb7bed896f0887df696601635f046adce38e2d7ba7: Status 404 returned error can't find the container with id f3ef61765a27b583a52d7ccb7bed896f0887df696601635f046adce38e2d7ba7 Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.783352 4706 generic.go:334] "Generic (PLEG): container finished" podID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerID="aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347" exitCode=0 Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.783526 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zsdd9" event={"ID":"63da40ba-a150-4396-bb8d-d8e4096a3961","Type":"ContainerDied","Data":"aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347"} Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.797384 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwksc" event={"ID":"32aaa66c-bc2d-42e3-9c17-681d58e0ef08","Type":"ContainerStarted","Data":"f3ef61765a27b583a52d7ccb7bed896f0887df696601635f046adce38e2d7ba7"} Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.836221 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.836299 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:24:35 crc kubenswrapper[4706]: I1208 19:24:35.926420 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.357111 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:36 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:36 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:36 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.357750 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.832971 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.834356 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.838899 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.843626 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.843665 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.856466 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5c785a58-21ca-47e0-8e0f-3b4576b3f300","Type":"ContainerStarted","Data":"05bad63074bfa8bfb843aa36741992024bed1a4a6df09dd77f541e4d59aa4956"} Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.856533 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5c785a58-21ca-47e0-8e0f-3b4576b3f300","Type":"ContainerStarted","Data":"287b75a882b59e820354ebd3ed8d4ec77d05c5252bf1138f73412a02455fb379"} Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.867534 4706 generic.go:334] "Generic (PLEG): container finished" podID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerID="374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d" exitCode=0 Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.867592 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwksc" event={"ID":"32aaa66c-bc2d-42e3-9c17-681d58e0ef08","Type":"ContainerDied","Data":"374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d"} Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.876848 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.8768263149999997 podStartE2EDuration="2.876826315s" podCreationTimestamp="2025-12-08 19:24:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:36.876293799 +0000 UTC m=+159.518494802" watchObservedRunningTime="2025-12-08 19:24:36.876826315 +0000 UTC m=+159.519027318" Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.941031 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:36 crc kubenswrapper[4706]: I1208 19:24:36.941382 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.043111 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.043196 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.043834 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.065952 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.122009 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.166407 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.353602 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:37 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:37 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:37 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.354583 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.573997 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 08 19:24:37 crc kubenswrapper[4706]: W1208 19:24:37.622391 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7bbab2b3_184f_4ef0_b98a_204aaa2c40ec.slice/crio-aab5ae03112c23591a7ae0c50d6b9123ad5cc779ef40f969cc1ad557e0cceaf9 WatchSource:0}: Error finding container aab5ae03112c23591a7ae0c50d6b9123ad5cc779ef40f969cc1ad557e0cceaf9: Status 404 returned error can't find the container with id aab5ae03112c23591a7ae0c50d6b9123ad5cc779ef40f969cc1ad557e0cceaf9 Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.903364 4706 generic.go:334] "Generic (PLEG): container finished" podID="5c785a58-21ca-47e0-8e0f-3b4576b3f300" containerID="05bad63074bfa8bfb843aa36741992024bed1a4a6df09dd77f541e4d59aa4956" exitCode=0 Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.903768 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5c785a58-21ca-47e0-8e0f-3b4576b3f300","Type":"ContainerDied","Data":"05bad63074bfa8bfb843aa36741992024bed1a4a6df09dd77f541e4d59aa4956"} Dec 08 19:24:37 crc kubenswrapper[4706]: I1208 19:24:37.918427 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec","Type":"ContainerStarted","Data":"aab5ae03112c23591a7ae0c50d6b9123ad5cc779ef40f969cc1ad557e0cceaf9"} Dec 08 19:24:38 crc kubenswrapper[4706]: I1208 19:24:38.353457 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:38 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:38 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:38 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:38 crc kubenswrapper[4706]: I1208 19:24:38.353529 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:38 crc kubenswrapper[4706]: I1208 19:24:38.954474 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec","Type":"ContainerStarted","Data":"812b32429496413d1be1a873c288abd6e6bb2c1197d1a8b20ff61c00c78dcbbf"} Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.227571 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.233305 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6rqvg" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.358651 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:39 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:39 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:39 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.358770 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.658471 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.742341 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kubelet-dir\") pod \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\" (UID: \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\") " Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.742426 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kube-api-access\") pod \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\" (UID: \"5c785a58-21ca-47e0-8e0f-3b4576b3f300\") " Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.742663 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5c785a58-21ca-47e0-8e0f-3b4576b3f300" (UID: "5c785a58-21ca-47e0-8e0f-3b4576b3f300"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.743362 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.752424 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5c785a58-21ca-47e0-8e0f-3b4576b3f300" (UID: "5c785a58-21ca-47e0-8e0f-3b4576b3f300"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.846433 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c785a58-21ca-47e0-8e0f-3b4576b3f300-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.997914 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.998581 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5c785a58-21ca-47e0-8e0f-3b4576b3f300","Type":"ContainerDied","Data":"287b75a882b59e820354ebd3ed8d4ec77d05c5252bf1138f73412a02455fb379"} Dec 08 19:24:39 crc kubenswrapper[4706]: I1208 19:24:39.998657 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="287b75a882b59e820354ebd3ed8d4ec77d05c5252bf1138f73412a02455fb379" Dec 08 19:24:40 crc kubenswrapper[4706]: I1208 19:24:40.030374 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.030351437 podStartE2EDuration="4.030351437s" podCreationTimestamp="2025-12-08 19:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:24:40.028949265 +0000 UTC m=+162.671150268" watchObservedRunningTime="2025-12-08 19:24:40.030351437 +0000 UTC m=+162.672552440" Dec 08 19:24:40 crc kubenswrapper[4706]: I1208 19:24:40.095894 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-ktnbl" Dec 08 19:24:40 crc kubenswrapper[4706]: I1208 19:24:40.350541 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:40 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:40 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:40 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:40 crc kubenswrapper[4706]: I1208 19:24:40.350624 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:41 crc kubenswrapper[4706]: I1208 19:24:41.010916 4706 generic.go:334] "Generic (PLEG): container finished" podID="7bbab2b3-184f-4ef0-b98a-204aaa2c40ec" containerID="812b32429496413d1be1a873c288abd6e6bb2c1197d1a8b20ff61c00c78dcbbf" exitCode=0 Dec 08 19:24:41 crc kubenswrapper[4706]: I1208 19:24:41.011326 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec","Type":"ContainerDied","Data":"812b32429496413d1be1a873c288abd6e6bb2c1197d1a8b20ff61c00c78dcbbf"} Dec 08 19:24:41 crc kubenswrapper[4706]: I1208 19:24:41.351356 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:41 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:41 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:41 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:41 crc kubenswrapper[4706]: I1208 19:24:41.351442 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:42 crc kubenswrapper[4706]: I1208 19:24:42.351025 4706 patch_prober.go:28] interesting pod/router-default-5444994796-cvvxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 19:24:42 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 08 19:24:42 crc kubenswrapper[4706]: [+]process-running ok Dec 08 19:24:42 crc kubenswrapper[4706]: healthz check failed Dec 08 19:24:42 crc kubenswrapper[4706]: I1208 19:24:42.351113 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cvvxf" podUID="65a505d5-7286-4b1c-ad40-f2e83c1e39d2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:24:43 crc kubenswrapper[4706]: I1208 19:24:43.412437 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:43 crc kubenswrapper[4706]: I1208 19:24:43.415854 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-cvvxf" Dec 08 19:24:43 crc kubenswrapper[4706]: I1208 19:24:43.627731 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:43 crc kubenswrapper[4706]: I1208 19:24:43.651112 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/439dd590-d737-435b-80a7-01f75fe6b64d-metrics-certs\") pod \"network-metrics-daemon-strs6\" (UID: \"439dd590-d737-435b-80a7-01f75fe6b64d\") " pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:43 crc kubenswrapper[4706]: I1208 19:24:43.943957 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-strs6" Dec 08 19:24:44 crc kubenswrapper[4706]: I1208 19:24:44.155446 4706 patch_prober.go:28] interesting pod/console-f9d7485db-zgngm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 08 19:24:44 crc kubenswrapper[4706]: I1208 19:24:44.155514 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zgngm" podUID="3f1e0eea-50be-4dd8-a504-dcdd049ae255" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 08 19:24:44 crc kubenswrapper[4706]: I1208 19:24:44.222371 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bxc4t" Dec 08 19:24:47 crc kubenswrapper[4706]: I1208 19:24:47.658972 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:47 crc kubenswrapper[4706]: I1208 19:24:47.824279 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kube-api-access\") pod \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\" (UID: \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\") " Dec 08 19:24:47 crc kubenswrapper[4706]: I1208 19:24:47.824402 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kubelet-dir\") pod \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\" (UID: \"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec\") " Dec 08 19:24:47 crc kubenswrapper[4706]: I1208 19:24:47.825887 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7bbab2b3-184f-4ef0-b98a-204aaa2c40ec" (UID: "7bbab2b3-184f-4ef0-b98a-204aaa2c40ec"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:24:47 crc kubenswrapper[4706]: I1208 19:24:47.826283 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:24:47 crc kubenswrapper[4706]: I1208 19:24:47.844389 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7bbab2b3-184f-4ef0-b98a-204aaa2c40ec" (UID: "7bbab2b3-184f-4ef0-b98a-204aaa2c40ec"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:24:47 crc kubenswrapper[4706]: I1208 19:24:47.928102 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bbab2b3-184f-4ef0-b98a-204aaa2c40ec-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 19:24:48 crc kubenswrapper[4706]: I1208 19:24:48.077150 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7bbab2b3-184f-4ef0-b98a-204aaa2c40ec","Type":"ContainerDied","Data":"aab5ae03112c23591a7ae0c50d6b9123ad5cc779ef40f969cc1ad557e0cceaf9"} Dec 08 19:24:48 crc kubenswrapper[4706]: I1208 19:24:48.077205 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aab5ae03112c23591a7ae0c50d6b9123ad5cc779ef40f969cc1ad557e0cceaf9" Dec 08 19:24:48 crc kubenswrapper[4706]: I1208 19:24:48.077284 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 19:24:52 crc kubenswrapper[4706]: I1208 19:24:52.990650 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:24:54 crc kubenswrapper[4706]: I1208 19:24:54.159364 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:24:54 crc kubenswrapper[4706]: I1208 19:24:54.168698 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:25:03 crc kubenswrapper[4706]: E1208 19:25:03.115000 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 08 19:25:03 crc kubenswrapper[4706]: E1208 19:25:03.115657 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5w7m4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-q6t9t_openshift-marketplace(9752bb6b-117b-4e2b-8794-a91e14dbeaf7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:25:03 crc kubenswrapper[4706]: E1208 19:25:03.119063 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-q6t9t" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" Dec 08 19:25:04 crc kubenswrapper[4706]: E1208 19:25:04.493539 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-q6t9t" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" Dec 08 19:25:04 crc kubenswrapper[4706]: E1208 19:25:04.559897 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 08 19:25:04 crc kubenswrapper[4706]: E1208 19:25:04.560102 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrvzb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zcrgl_openshift-marketplace(0415ede8-d911-4058-ba7f-507f756ef909): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:25:04 crc kubenswrapper[4706]: E1208 19:25:04.561352 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-zcrgl" podUID="0415ede8-d911-4058-ba7f-507f756ef909" Dec 08 19:25:04 crc kubenswrapper[4706]: I1208 19:25:04.857868 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 19:25:05 crc kubenswrapper[4706]: I1208 19:25:05.019317 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n2bsv" Dec 08 19:25:05 crc kubenswrapper[4706]: I1208 19:25:05.835931 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:25:05 crc kubenswrapper[4706]: I1208 19:25:05.836011 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:25:08 crc kubenswrapper[4706]: E1208 19:25:08.557206 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 08 19:25:08 crc kubenswrapper[4706]: E1208 19:25:08.558171 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fsnb5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wwksc_openshift-marketplace(32aaa66c-bc2d-42e3-9c17-681d58e0ef08): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:25:08 crc kubenswrapper[4706]: E1208 19:25:08.559440 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wwksc" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.097934 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-wwksc" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.173543 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.173923 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-whp92,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6fmnd_openshift-marketplace(26f226b0-fa6c-4efc-af99-f69c1ad8a464): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.175085 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6fmnd" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.188506 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.188794 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mxkgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2h6vv_openshift-marketplace(443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.194388 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2h6vv" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.217451 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.218021 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lxfrp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zsdd9_openshift-marketplace(63da40ba-a150-4396-bb8d-d8e4096a3961): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.219382 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zsdd9" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.230481 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2h6vv" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.230972 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6fmnd" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.280283 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.280875 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4rgbr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-njntc_openshift-marketplace(0a9bae31-9b10-44fb-b30f-7cfca35cda15): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.282982 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-njntc" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.284461 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.284621 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d5pxf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-8rl9w_openshift-marketplace(3200b927-522a-4762-a731-c56780f4a137): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.285883 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-8rl9w" podUID="3200b927-522a-4762-a731-c56780f4a137" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.423371 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.423837 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c785a58-21ca-47e0-8e0f-3b4576b3f300" containerName="pruner" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.423859 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c785a58-21ca-47e0-8e0f-3b4576b3f300" containerName="pruner" Dec 08 19:25:10 crc kubenswrapper[4706]: E1208 19:25:10.423882 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bbab2b3-184f-4ef0-b98a-204aaa2c40ec" containerName="pruner" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.423891 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bbab2b3-184f-4ef0-b98a-204aaa2c40ec" containerName="pruner" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.424034 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bbab2b3-184f-4ef0-b98a-204aaa2c40ec" containerName="pruner" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.424066 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c785a58-21ca-47e0-8e0f-3b4576b3f300" containerName="pruner" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.424710 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.427808 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.430066 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.433409 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.516242 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-strs6"] Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.600536 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df3f704-6b34-42a0-b531-2472aed802ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4df3f704-6b34-42a0-b531-2472aed802ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.600659 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df3f704-6b34-42a0-b531-2472aed802ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4df3f704-6b34-42a0-b531-2472aed802ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.701861 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df3f704-6b34-42a0-b531-2472aed802ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4df3f704-6b34-42a0-b531-2472aed802ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.702123 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df3f704-6b34-42a0-b531-2472aed802ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4df3f704-6b34-42a0-b531-2472aed802ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.702004 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df3f704-6b34-42a0-b531-2472aed802ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4df3f704-6b34-42a0-b531-2472aed802ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.726371 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df3f704-6b34-42a0-b531-2472aed802ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4df3f704-6b34-42a0-b531-2472aed802ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:10 crc kubenswrapper[4706]: I1208 19:25:10.759888 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:11 crc kubenswrapper[4706]: I1208 19:25:11.177950 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 08 19:25:11 crc kubenswrapper[4706]: I1208 19:25:11.235163 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4df3f704-6b34-42a0-b531-2472aed802ff","Type":"ContainerStarted","Data":"807daf13a8481588aa39ac991bde7a3c716889cee1e6e0b8a7210dce51aad533"} Dec 08 19:25:11 crc kubenswrapper[4706]: I1208 19:25:11.240804 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-strs6" event={"ID":"439dd590-d737-435b-80a7-01f75fe6b64d","Type":"ContainerStarted","Data":"9fdce428659ac761d15d77263b51c3299b68474ac5770232557945d7fb70f700"} Dec 08 19:25:11 crc kubenswrapper[4706]: I1208 19:25:11.240830 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-strs6" event={"ID":"439dd590-d737-435b-80a7-01f75fe6b64d","Type":"ContainerStarted","Data":"ea3c3111660f013c558bb9c8f198a2ab62548673f74fd9810ed2da5616325805"} Dec 08 19:25:11 crc kubenswrapper[4706]: I1208 19:25:11.241007 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-strs6" event={"ID":"439dd590-d737-435b-80a7-01f75fe6b64d","Type":"ContainerStarted","Data":"aa8fa42f05f38d70eee8db9dab09621e23fe617ca92e43623a5586934ae0a8bb"} Dec 08 19:25:11 crc kubenswrapper[4706]: E1208 19:25:11.243650 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-njntc" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" Dec 08 19:25:11 crc kubenswrapper[4706]: E1208 19:25:11.244206 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zsdd9" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" Dec 08 19:25:11 crc kubenswrapper[4706]: E1208 19:25:11.250757 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-8rl9w" podUID="3200b927-522a-4762-a731-c56780f4a137" Dec 08 19:25:11 crc kubenswrapper[4706]: I1208 19:25:11.259725 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-strs6" podStartSLOduration=170.259696497 podStartE2EDuration="2m50.259696497s" podCreationTimestamp="2025-12-08 19:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:25:11.256229791 +0000 UTC m=+193.898430804" watchObservedRunningTime="2025-12-08 19:25:11.259696497 +0000 UTC m=+193.901897500" Dec 08 19:25:12 crc kubenswrapper[4706]: I1208 19:25:12.258867 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4df3f704-6b34-42a0-b531-2472aed802ff","Type":"ContainerStarted","Data":"315a1c0b5e74722f3a567292d05f7887a5b4cbc017256d9305d39e50337d525c"} Dec 08 19:25:12 crc kubenswrapper[4706]: I1208 19:25:12.278603 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.2785828 podStartE2EDuration="2.2785828s" podCreationTimestamp="2025-12-08 19:25:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:25:12.273414231 +0000 UTC m=+194.915615234" watchObservedRunningTime="2025-12-08 19:25:12.2785828 +0000 UTC m=+194.920783803" Dec 08 19:25:12 crc kubenswrapper[4706]: I1208 19:25:12.872002 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq5pd"] Dec 08 19:25:13 crc kubenswrapper[4706]: I1208 19:25:13.268922 4706 generic.go:334] "Generic (PLEG): container finished" podID="4df3f704-6b34-42a0-b531-2472aed802ff" containerID="315a1c0b5e74722f3a567292d05f7887a5b4cbc017256d9305d39e50337d525c" exitCode=0 Dec 08 19:25:13 crc kubenswrapper[4706]: I1208 19:25:13.269001 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4df3f704-6b34-42a0-b531-2472aed802ff","Type":"ContainerDied","Data":"315a1c0b5e74722f3a567292d05f7887a5b4cbc017256d9305d39e50337d525c"} Dec 08 19:25:14 crc kubenswrapper[4706]: I1208 19:25:14.603621 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:14 crc kubenswrapper[4706]: I1208 19:25:14.758360 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4df3f704-6b34-42a0-b531-2472aed802ff-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4df3f704-6b34-42a0-b531-2472aed802ff" (UID: "4df3f704-6b34-42a0-b531-2472aed802ff"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:25:14 crc kubenswrapper[4706]: I1208 19:25:14.758447 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df3f704-6b34-42a0-b531-2472aed802ff-kubelet-dir\") pod \"4df3f704-6b34-42a0-b531-2472aed802ff\" (UID: \"4df3f704-6b34-42a0-b531-2472aed802ff\") " Dec 08 19:25:14 crc kubenswrapper[4706]: I1208 19:25:14.758503 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df3f704-6b34-42a0-b531-2472aed802ff-kube-api-access\") pod \"4df3f704-6b34-42a0-b531-2472aed802ff\" (UID: \"4df3f704-6b34-42a0-b531-2472aed802ff\") " Dec 08 19:25:14 crc kubenswrapper[4706]: I1208 19:25:14.759962 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df3f704-6b34-42a0-b531-2472aed802ff-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:14 crc kubenswrapper[4706]: I1208 19:25:14.765076 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4df3f704-6b34-42a0-b531-2472aed802ff-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4df3f704-6b34-42a0-b531-2472aed802ff" (UID: "4df3f704-6b34-42a0-b531-2472aed802ff"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:14 crc kubenswrapper[4706]: I1208 19:25:14.862191 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df3f704-6b34-42a0-b531-2472aed802ff-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:15 crc kubenswrapper[4706]: I1208 19:25:15.282100 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4df3f704-6b34-42a0-b531-2472aed802ff","Type":"ContainerDied","Data":"807daf13a8481588aa39ac991bde7a3c716889cee1e6e0b8a7210dce51aad533"} Dec 08 19:25:15 crc kubenswrapper[4706]: I1208 19:25:15.282155 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="807daf13a8481588aa39ac991bde7a3c716889cee1e6e0b8a7210dce51aad533" Dec 08 19:25:15 crc kubenswrapper[4706]: I1208 19:25:15.282204 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.406738 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 08 19:25:16 crc kubenswrapper[4706]: E1208 19:25:16.407345 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4df3f704-6b34-42a0-b531-2472aed802ff" containerName="pruner" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.407360 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df3f704-6b34-42a0-b531-2472aed802ff" containerName="pruner" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.407487 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4df3f704-6b34-42a0-b531-2472aed802ff" containerName="pruner" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.407946 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.411220 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.411317 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.419188 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.488779 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-var-lock\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.488858 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d87b9a0-63a8-4619-88cb-127f2608edd6-kube-api-access\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.489030 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.590683 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-var-lock\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.590800 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d87b9a0-63a8-4619-88cb-127f2608edd6-kube-api-access\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.590842 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-var-lock\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.590872 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.590950 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.608456 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d87b9a0-63a8-4619-88cb-127f2608edd6-kube-api-access\") pod \"installer-9-crc\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:16 crc kubenswrapper[4706]: I1208 19:25:16.738920 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:17 crc kubenswrapper[4706]: I1208 19:25:17.160610 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 08 19:25:17 crc kubenswrapper[4706]: I1208 19:25:17.303118 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1d87b9a0-63a8-4619-88cb-127f2608edd6","Type":"ContainerStarted","Data":"768d2e67ef4da395b788b1b59293723eb42bef1bd920542cd84f2d0e6b788624"} Dec 08 19:25:18 crc kubenswrapper[4706]: I1208 19:25:18.319329 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1d87b9a0-63a8-4619-88cb-127f2608edd6","Type":"ContainerStarted","Data":"60df5744296e422b2735860df0b42e8d7235bac12190d4f93629d5ea63a13a68"} Dec 08 19:25:18 crc kubenswrapper[4706]: I1208 19:25:18.322191 4706 generic.go:334] "Generic (PLEG): container finished" podID="0415ede8-d911-4058-ba7f-507f756ef909" containerID="562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a" exitCode=0 Dec 08 19:25:18 crc kubenswrapper[4706]: I1208 19:25:18.322228 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zcrgl" event={"ID":"0415ede8-d911-4058-ba7f-507f756ef909","Type":"ContainerDied","Data":"562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a"} Dec 08 19:25:18 crc kubenswrapper[4706]: I1208 19:25:18.346282 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.346232509 podStartE2EDuration="2.346232509s" podCreationTimestamp="2025-12-08 19:25:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:25:18.34437609 +0000 UTC m=+200.986577093" watchObservedRunningTime="2025-12-08 19:25:18.346232509 +0000 UTC m=+200.988433522" Dec 08 19:25:19 crc kubenswrapper[4706]: I1208 19:25:19.331004 4706 generic.go:334] "Generic (PLEG): container finished" podID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerID="06466ad7b3b1602833841a787d961f883ae924716ae11da2976838092631ea05" exitCode=0 Dec 08 19:25:19 crc kubenswrapper[4706]: I1208 19:25:19.331108 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6t9t" event={"ID":"9752bb6b-117b-4e2b-8794-a91e14dbeaf7","Type":"ContainerDied","Data":"06466ad7b3b1602833841a787d961f883ae924716ae11da2976838092631ea05"} Dec 08 19:25:19 crc kubenswrapper[4706]: I1208 19:25:19.337111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zcrgl" event={"ID":"0415ede8-d911-4058-ba7f-507f756ef909","Type":"ContainerStarted","Data":"f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291"} Dec 08 19:25:19 crc kubenswrapper[4706]: I1208 19:25:19.384968 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zcrgl" podStartSLOduration=3.338466088 podStartE2EDuration="46.384942965s" podCreationTimestamp="2025-12-08 19:24:33 +0000 UTC" firstStartedPulling="2025-12-08 19:24:35.811650145 +0000 UTC m=+158.453851148" lastFinishedPulling="2025-12-08 19:25:18.858127012 +0000 UTC m=+201.500328025" observedRunningTime="2025-12-08 19:25:19.383091497 +0000 UTC m=+202.025292510" watchObservedRunningTime="2025-12-08 19:25:19.384942965 +0000 UTC m=+202.027143968" Dec 08 19:25:20 crc kubenswrapper[4706]: I1208 19:25:20.343381 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6t9t" event={"ID":"9752bb6b-117b-4e2b-8794-a91e14dbeaf7","Type":"ContainerStarted","Data":"9b8b6850d6b092bc69c7a6b89e103774a305bfe007226cea7b138b5813292af3"} Dec 08 19:25:20 crc kubenswrapper[4706]: I1208 19:25:20.362469 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q6t9t" podStartSLOduration=4.223064638 podStartE2EDuration="50.362444723s" podCreationTimestamp="2025-12-08 19:24:30 +0000 UTC" firstStartedPulling="2025-12-08 19:24:33.602613526 +0000 UTC m=+156.244814529" lastFinishedPulling="2025-12-08 19:25:19.741993611 +0000 UTC m=+202.384194614" observedRunningTime="2025-12-08 19:25:20.361219664 +0000 UTC m=+203.003420667" watchObservedRunningTime="2025-12-08 19:25:20.362444723 +0000 UTC m=+203.004645726" Dec 08 19:25:21 crc kubenswrapper[4706]: I1208 19:25:21.358073 4706 generic.go:334] "Generic (PLEG): container finished" podID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerID="d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494" exitCode=0 Dec 08 19:25:21 crc kubenswrapper[4706]: I1208 19:25:21.358247 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fmnd" event={"ID":"26f226b0-fa6c-4efc-af99-f69c1ad8a464","Type":"ContainerDied","Data":"d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494"} Dec 08 19:25:22 crc kubenswrapper[4706]: I1208 19:25:22.162479 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:25:22 crc kubenswrapper[4706]: I1208 19:25:22.162566 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:25:22 crc kubenswrapper[4706]: I1208 19:25:22.249626 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:25:22 crc kubenswrapper[4706]: I1208 19:25:22.367059 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fmnd" event={"ID":"26f226b0-fa6c-4efc-af99-f69c1ad8a464","Type":"ContainerStarted","Data":"6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764"} Dec 08 19:25:22 crc kubenswrapper[4706]: I1208 19:25:22.369580 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h6vv" event={"ID":"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5","Type":"ContainerStarted","Data":"689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c"} Dec 08 19:25:22 crc kubenswrapper[4706]: I1208 19:25:22.390215 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6fmnd" podStartSLOduration=3.403403104 podStartE2EDuration="50.39018445s" podCreationTimestamp="2025-12-08 19:24:32 +0000 UTC" firstStartedPulling="2025-12-08 19:24:34.760541626 +0000 UTC m=+157.402742629" lastFinishedPulling="2025-12-08 19:25:21.747322972 +0000 UTC m=+204.389523975" observedRunningTime="2025-12-08 19:25:22.389972594 +0000 UTC m=+205.032173597" watchObservedRunningTime="2025-12-08 19:25:22.39018445 +0000 UTC m=+205.032385473" Dec 08 19:25:23 crc kubenswrapper[4706]: I1208 19:25:23.164841 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:25:23 crc kubenswrapper[4706]: I1208 19:25:23.165413 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:25:23 crc kubenswrapper[4706]: I1208 19:25:23.379711 4706 generic.go:334] "Generic (PLEG): container finished" podID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerID="689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c" exitCode=0 Dec 08 19:25:23 crc kubenswrapper[4706]: I1208 19:25:23.379786 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h6vv" event={"ID":"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5","Type":"ContainerDied","Data":"689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c"} Dec 08 19:25:23 crc kubenswrapper[4706]: I1208 19:25:23.553374 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:25:23 crc kubenswrapper[4706]: I1208 19:25:23.553458 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:25:23 crc kubenswrapper[4706]: I1208 19:25:23.600435 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:25:24 crc kubenswrapper[4706]: I1208 19:25:24.203822 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-6fmnd" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="registry-server" probeResult="failure" output=< Dec 08 19:25:24 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:25:24 crc kubenswrapper[4706]: > Dec 08 19:25:24 crc kubenswrapper[4706]: I1208 19:25:24.440816 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:25:27 crc kubenswrapper[4706]: I1208 19:25:27.872992 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zcrgl"] Dec 08 19:25:27 crc kubenswrapper[4706]: I1208 19:25:27.873521 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zcrgl" podUID="0415ede8-d911-4058-ba7f-507f756ef909" containerName="registry-server" containerID="cri-o://f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291" gracePeriod=2 Dec 08 19:25:28 crc kubenswrapper[4706]: I1208 19:25:28.421386 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h6vv" event={"ID":"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5","Type":"ContainerStarted","Data":"09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae"} Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.381235 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.435765 4706 generic.go:334] "Generic (PLEG): container finished" podID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerID="25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee" exitCode=0 Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.435841 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-njntc" event={"ID":"0a9bae31-9b10-44fb-b30f-7cfca35cda15","Type":"ContainerDied","Data":"25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee"} Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.444068 4706 generic.go:334] "Generic (PLEG): container finished" podID="0415ede8-d911-4058-ba7f-507f756ef909" containerID="f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291" exitCode=0 Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.444169 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zcrgl" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.444312 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zcrgl" event={"ID":"0415ede8-d911-4058-ba7f-507f756ef909","Type":"ContainerDied","Data":"f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291"} Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.444388 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zcrgl" event={"ID":"0415ede8-d911-4058-ba7f-507f756ef909","Type":"ContainerDied","Data":"1a4ba1c80c852d76fa79a53cbd27523737c25ea28c411e09e8206f17e36b9e20"} Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.444422 4706 scope.go:117] "RemoveContainer" containerID="f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.447997 4706 generic.go:334] "Generic (PLEG): container finished" podID="3200b927-522a-4762-a731-c56780f4a137" containerID="f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151" exitCode=0 Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.448785 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rl9w" event={"ID":"3200b927-522a-4762-a731-c56780f4a137","Type":"ContainerDied","Data":"f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151"} Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.489297 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2h6vv" podStartSLOduration=4.558683057 podStartE2EDuration="59.489250798s" podCreationTimestamp="2025-12-08 19:24:30 +0000 UTC" firstStartedPulling="2025-12-08 19:24:32.43858574 +0000 UTC m=+155.080786743" lastFinishedPulling="2025-12-08 19:25:27.369153481 +0000 UTC m=+210.011354484" observedRunningTime="2025-12-08 19:25:29.476238148 +0000 UTC m=+212.118439171" watchObservedRunningTime="2025-12-08 19:25:29.489250798 +0000 UTC m=+212.131451801" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.490780 4706 scope.go:117] "RemoveContainer" containerID="562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.518401 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-utilities\") pod \"0415ede8-d911-4058-ba7f-507f756ef909\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.518582 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrvzb\" (UniqueName: \"kubernetes.io/projected/0415ede8-d911-4058-ba7f-507f756ef909-kube-api-access-jrvzb\") pod \"0415ede8-d911-4058-ba7f-507f756ef909\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.518615 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-catalog-content\") pod \"0415ede8-d911-4058-ba7f-507f756ef909\" (UID: \"0415ede8-d911-4058-ba7f-507f756ef909\") " Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.520006 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-utilities" (OuterVolumeSpecName: "utilities") pod "0415ede8-d911-4058-ba7f-507f756ef909" (UID: "0415ede8-d911-4058-ba7f-507f756ef909"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.524659 4706 scope.go:117] "RemoveContainer" containerID="c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.528181 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0415ede8-d911-4058-ba7f-507f756ef909-kube-api-access-jrvzb" (OuterVolumeSpecName: "kube-api-access-jrvzb") pod "0415ede8-d911-4058-ba7f-507f756ef909" (UID: "0415ede8-d911-4058-ba7f-507f756ef909"). InnerVolumeSpecName "kube-api-access-jrvzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.541953 4706 scope.go:117] "RemoveContainer" containerID="f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291" Dec 08 19:25:29 crc kubenswrapper[4706]: E1208 19:25:29.542646 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291\": container with ID starting with f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291 not found: ID does not exist" containerID="f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.542719 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291"} err="failed to get container status \"f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291\": rpc error: code = NotFound desc = could not find container \"f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291\": container with ID starting with f67a28a78b4131259d11ebb078deee89b0bfd782577c82e82da7018d65c54291 not found: ID does not exist" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.542802 4706 scope.go:117] "RemoveContainer" containerID="562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a" Dec 08 19:25:29 crc kubenswrapper[4706]: E1208 19:25:29.543962 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a\": container with ID starting with 562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a not found: ID does not exist" containerID="562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.544019 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a"} err="failed to get container status \"562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a\": rpc error: code = NotFound desc = could not find container \"562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a\": container with ID starting with 562e08eb191d525da0532db116dabf15c1ad12ade37a8b6cff460728438fec1a not found: ID does not exist" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.544049 4706 scope.go:117] "RemoveContainer" containerID="c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4" Dec 08 19:25:29 crc kubenswrapper[4706]: E1208 19:25:29.545322 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4\": container with ID starting with c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4 not found: ID does not exist" containerID="c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.545348 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4"} err="failed to get container status \"c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4\": rpc error: code = NotFound desc = could not find container \"c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4\": container with ID starting with c9eea714479b67b46b106c8f95036ae7f5b2488e71f932ad51579f9b978418d4 not found: ID does not exist" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.549339 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0415ede8-d911-4058-ba7f-507f756ef909" (UID: "0415ede8-d911-4058-ba7f-507f756ef909"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.620725 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.620758 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrvzb\" (UniqueName: \"kubernetes.io/projected/0415ede8-d911-4058-ba7f-507f756ef909-kube-api-access-jrvzb\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.620768 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0415ede8-d911-4058-ba7f-507f756ef909-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.782441 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zcrgl"] Dec 08 19:25:29 crc kubenswrapper[4706]: I1208 19:25:29.787621 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zcrgl"] Dec 08 19:25:30 crc kubenswrapper[4706]: I1208 19:25:30.460181 4706 generic.go:334] "Generic (PLEG): container finished" podID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerID="6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13" exitCode=0 Dec 08 19:25:30 crc kubenswrapper[4706]: I1208 19:25:30.460295 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwksc" event={"ID":"32aaa66c-bc2d-42e3-9c17-681d58e0ef08","Type":"ContainerDied","Data":"6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13"} Dec 08 19:25:30 crc kubenswrapper[4706]: I1208 19:25:30.464558 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rl9w" event={"ID":"3200b927-522a-4762-a731-c56780f4a137","Type":"ContainerStarted","Data":"eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c"} Dec 08 19:25:30 crc kubenswrapper[4706]: I1208 19:25:30.467590 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-njntc" event={"ID":"0a9bae31-9b10-44fb-b30f-7cfca35cda15","Type":"ContainerStarted","Data":"ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e"} Dec 08 19:25:30 crc kubenswrapper[4706]: I1208 19:25:30.471812 4706 generic.go:334] "Generic (PLEG): container finished" podID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerID="9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840" exitCode=0 Dec 08 19:25:30 crc kubenswrapper[4706]: I1208 19:25:30.471867 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zsdd9" event={"ID":"63da40ba-a150-4396-bb8d-d8e4096a3961","Type":"ContainerDied","Data":"9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840"} Dec 08 19:25:30 crc kubenswrapper[4706]: I1208 19:25:30.505553 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-njntc" podStartSLOduration=3.104298013 podStartE2EDuration="1m0.505531828s" podCreationTimestamp="2025-12-08 19:24:30 +0000 UTC" firstStartedPulling="2025-12-08 19:24:32.466981229 +0000 UTC m=+155.109182232" lastFinishedPulling="2025-12-08 19:25:29.868215044 +0000 UTC m=+212.510416047" observedRunningTime="2025-12-08 19:25:30.503881236 +0000 UTC m=+213.146082249" watchObservedRunningTime="2025-12-08 19:25:30.505531828 +0000 UTC m=+213.147732831" Dec 08 19:25:30 crc kubenswrapper[4706]: I1208 19:25:30.527236 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8rl9w" podStartSLOduration=3.069236784 podStartE2EDuration="59.527207221s" podCreationTimestamp="2025-12-08 19:24:31 +0000 UTC" firstStartedPulling="2025-12-08 19:24:33.602607786 +0000 UTC m=+156.244808799" lastFinishedPulling="2025-12-08 19:25:30.060578233 +0000 UTC m=+212.702779236" observedRunningTime="2025-12-08 19:25:30.522150112 +0000 UTC m=+213.164351135" watchObservedRunningTime="2025-12-08 19:25:30.527207221 +0000 UTC m=+213.169408224" Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.055126 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.055524 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.126439 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.455536 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.459232 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.477531 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zsdd9" event={"ID":"63da40ba-a150-4396-bb8d-d8e4096a3961","Type":"ContainerStarted","Data":"fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969"} Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.482599 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwksc" event={"ID":"32aaa66c-bc2d-42e3-9c17-681d58e0ef08","Type":"ContainerStarted","Data":"cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926"} Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.512400 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zsdd9" podStartSLOduration=3.378825803 podStartE2EDuration="58.51237806s" podCreationTimestamp="2025-12-08 19:24:33 +0000 UTC" firstStartedPulling="2025-12-08 19:24:35.785559927 +0000 UTC m=+158.427760930" lastFinishedPulling="2025-12-08 19:25:30.919112184 +0000 UTC m=+213.561313187" observedRunningTime="2025-12-08 19:25:31.509389246 +0000 UTC m=+214.151590249" watchObservedRunningTime="2025-12-08 19:25:31.51237806 +0000 UTC m=+214.154579053" Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.532267 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wwksc" podStartSLOduration=3.412758612 podStartE2EDuration="57.532028599s" podCreationTimestamp="2025-12-08 19:24:34 +0000 UTC" firstStartedPulling="2025-12-08 19:24:36.869611544 +0000 UTC m=+159.511812557" lastFinishedPulling="2025-12-08 19:25:30.988881541 +0000 UTC m=+213.631082544" observedRunningTime="2025-12-08 19:25:31.52922218 +0000 UTC m=+214.171423183" watchObservedRunningTime="2025-12-08 19:25:31.532028599 +0000 UTC m=+214.174229612" Dec 08 19:25:31 crc kubenswrapper[4706]: I1208 19:25:31.616153 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0415ede8-d911-4058-ba7f-507f756ef909" path="/var/lib/kubelet/pods/0415ede8-d911-4058-ba7f-507f756ef909/volumes" Dec 08 19:25:32 crc kubenswrapper[4706]: I1208 19:25:32.218051 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:25:32 crc kubenswrapper[4706]: I1208 19:25:32.296066 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:25:32 crc kubenswrapper[4706]: I1208 19:25:32.296141 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:25:32 crc kubenswrapper[4706]: I1208 19:25:32.505494 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-njntc" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="registry-server" probeResult="failure" output=< Dec 08 19:25:32 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:25:32 crc kubenswrapper[4706]: > Dec 08 19:25:33 crc kubenswrapper[4706]: I1208 19:25:33.207776 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:25:33 crc kubenswrapper[4706]: I1208 19:25:33.249194 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:25:33 crc kubenswrapper[4706]: I1208 19:25:33.332874 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-8rl9w" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="registry-server" probeResult="failure" output=< Dec 08 19:25:33 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:25:33 crc kubenswrapper[4706]: > Dec 08 19:25:34 crc kubenswrapper[4706]: I1208 19:25:34.207606 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:25:34 crc kubenswrapper[4706]: I1208 19:25:34.208042 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:25:34 crc kubenswrapper[4706]: I1208 19:25:34.551184 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:25:34 crc kubenswrapper[4706]: I1208 19:25:34.551297 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:25:35 crc kubenswrapper[4706]: I1208 19:25:35.257708 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zsdd9" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="registry-server" probeResult="failure" output=< Dec 08 19:25:35 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:25:35 crc kubenswrapper[4706]: > Dec 08 19:25:35 crc kubenswrapper[4706]: I1208 19:25:35.588389 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wwksc" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="registry-server" probeResult="failure" output=< Dec 08 19:25:35 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:25:35 crc kubenswrapper[4706]: > Dec 08 19:25:35 crc kubenswrapper[4706]: I1208 19:25:35.836202 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:25:35 crc kubenswrapper[4706]: I1208 19:25:35.836317 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:25:35 crc kubenswrapper[4706]: I1208 19:25:35.836395 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:25:35 crc kubenswrapper[4706]: I1208 19:25:35.837283 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:25:35 crc kubenswrapper[4706]: I1208 19:25:35.837362 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12" gracePeriod=600 Dec 08 19:25:37 crc kubenswrapper[4706]: I1208 19:25:37.536619 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12" exitCode=0 Dec 08 19:25:37 crc kubenswrapper[4706]: I1208 19:25:37.536698 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12"} Dec 08 19:25:37 crc kubenswrapper[4706]: I1208 19:25:37.915828 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" podUID="c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" containerName="oauth-openshift" containerID="cri-o://4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b" gracePeriod=15 Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.332704 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.371906 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-74b487c797-zqwll"] Dec 08 19:25:38 crc kubenswrapper[4706]: E1208 19:25:38.372287 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0415ede8-d911-4058-ba7f-507f756ef909" containerName="extract-content" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.372307 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0415ede8-d911-4058-ba7f-507f756ef909" containerName="extract-content" Dec 08 19:25:38 crc kubenswrapper[4706]: E1208 19:25:38.372334 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0415ede8-d911-4058-ba7f-507f756ef909" containerName="registry-server" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.372341 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0415ede8-d911-4058-ba7f-507f756ef909" containerName="registry-server" Dec 08 19:25:38 crc kubenswrapper[4706]: E1208 19:25:38.372355 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" containerName="oauth-openshift" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.372363 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" containerName="oauth-openshift" Dec 08 19:25:38 crc kubenswrapper[4706]: E1208 19:25:38.372371 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0415ede8-d911-4058-ba7f-507f756ef909" containerName="extract-utilities" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.372379 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0415ede8-d911-4058-ba7f-507f756ef909" containerName="extract-utilities" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.372518 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0415ede8-d911-4058-ba7f-507f756ef909" containerName="registry-server" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.372531 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" containerName="oauth-openshift" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.373115 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.385323 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74b487c797-zqwll"] Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.464542 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-trusted-ca-bundle\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.464597 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t68v8\" (UniqueName: \"kubernetes.io/projected/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-kube-api-access-t68v8\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.464649 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-error\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.464716 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-router-certs\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465215 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-login\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465387 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-idp-0-file-data\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465504 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-session\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465570 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-policies\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465640 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-serving-cert\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465695 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465731 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-dir\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465783 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-provider-selection\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465819 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-ocp-branding-template\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465831 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465856 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-cliconfig\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.465908 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-service-ca\") pod \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\" (UID: \"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf\") " Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466077 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466187 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-session\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466244 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466401 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466443 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-login\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466520 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466583 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466591 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-router-certs\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466623 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466651 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466696 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-audit-policies\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466732 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466773 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466916 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-error\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466951 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/feb82476-5d10-4697-b2ef-85e52229af07-audit-dir\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.466981 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9ftn\" (UniqueName: \"kubernetes.io/projected/feb82476-5d10-4697-b2ef-85e52229af07-kube-api-access-w9ftn\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.467006 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-service-ca\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.467195 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.467219 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.467235 4706 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.467247 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.467275 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.471253 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.472212 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.472241 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-kube-api-access-t68v8" (OuterVolumeSpecName: "kube-api-access-t68v8") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "kube-api-access-t68v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.472534 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.472766 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.472831 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.473209 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.473331 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.473790 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" (UID: "c14f56cb-dc94-4ccf-af56-cd61ec40c6bf"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.544996 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"9ab30946f44cee272e62060d2c6e0df51babab84dbf0a98d8b03bd4d916925ee"} Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.549381 4706 generic.go:334] "Generic (PLEG): container finished" podID="c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" containerID="4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b" exitCode=0 Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.549647 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" event={"ID":"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf","Type":"ContainerDied","Data":"4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b"} Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.549724 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" event={"ID":"c14f56cb-dc94-4ccf-af56-cd61ec40c6bf","Type":"ContainerDied","Data":"5cdcbb16d30e26197192ac37b3826639b4bdb8841290c8d2c84a3df8ccc5752c"} Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.549805 4706 scope.go:117] "RemoveContainer" containerID="4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.549984 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bq5pd" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.568819 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569106 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-router-certs\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569364 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-audit-policies\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569388 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569414 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569546 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-error\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569572 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/feb82476-5d10-4697-b2ef-85e52229af07-audit-dir\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569594 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-service-ca\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569611 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9ftn\" (UniqueName: \"kubernetes.io/projected/feb82476-5d10-4697-b2ef-85e52229af07-kube-api-access-w9ftn\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569647 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-session\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569667 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569718 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569741 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-login\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569824 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t68v8\" (UniqueName: \"kubernetes.io/projected/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-kube-api-access-t68v8\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.569838 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.570013 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.570976 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-audit-policies\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571243 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-service-ca\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571304 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/feb82476-5d10-4697-b2ef-85e52229af07-audit-dir\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571354 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571640 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571750 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571779 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571806 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571837 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.571862 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.572294 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.573050 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.573874 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-router-certs\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.574655 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-login\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.576045 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-session\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.576207 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.576527 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.577132 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-user-template-error\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.578137 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/feb82476-5d10-4697-b2ef-85e52229af07-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.587631 4706 scope.go:117] "RemoveContainer" containerID="4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b" Dec 08 19:25:38 crc kubenswrapper[4706]: E1208 19:25:38.588277 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b\": container with ID starting with 4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b not found: ID does not exist" containerID="4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.588329 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b"} err="failed to get container status \"4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b\": rpc error: code = NotFound desc = could not find container \"4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b\": container with ID starting with 4543049a4a299a1f781f535bec520624dd0119b57abb38c519240560b647fe5b not found: ID does not exist" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.592074 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9ftn\" (UniqueName: \"kubernetes.io/projected/feb82476-5d10-4697-b2ef-85e52229af07-kube-api-access-w9ftn\") pod \"oauth-openshift-74b487c797-zqwll\" (UID: \"feb82476-5d10-4697-b2ef-85e52229af07\") " pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.594713 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq5pd"] Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.600208 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq5pd"] Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.691927 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:38 crc kubenswrapper[4706]: I1208 19:25:38.921592 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74b487c797-zqwll"] Dec 08 19:25:39 crc kubenswrapper[4706]: I1208 19:25:39.559708 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" event={"ID":"feb82476-5d10-4697-b2ef-85e52229af07","Type":"ContainerStarted","Data":"037b5822b5a289a0dfa726e63abd1e10e8682bad9ca6c97ad0f192c402d8734b"} Dec 08 19:25:39 crc kubenswrapper[4706]: I1208 19:25:39.559764 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" event={"ID":"feb82476-5d10-4697-b2ef-85e52229af07","Type":"ContainerStarted","Data":"0c678213c83895f756775ceb56a65bad9afd0d697a1499a0b304f7a268028e8d"} Dec 08 19:25:39 crc kubenswrapper[4706]: I1208 19:25:39.585889 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" podStartSLOduration=27.58586916 podStartE2EDuration="27.58586916s" podCreationTimestamp="2025-12-08 19:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:25:39.581556194 +0000 UTC m=+222.223757217" watchObservedRunningTime="2025-12-08 19:25:39.58586916 +0000 UTC m=+222.228070163" Dec 08 19:25:39 crc kubenswrapper[4706]: I1208 19:25:39.617290 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c14f56cb-dc94-4ccf-af56-cd61ec40c6bf" path="/var/lib/kubelet/pods/c14f56cb-dc94-4ccf-af56-cd61ec40c6bf/volumes" Dec 08 19:25:40 crc kubenswrapper[4706]: I1208 19:25:40.568203 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:40 crc kubenswrapper[4706]: I1208 19:25:40.575492 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-74b487c797-zqwll" Dec 08 19:25:41 crc kubenswrapper[4706]: I1208 19:25:41.100742 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:25:41 crc kubenswrapper[4706]: I1208 19:25:41.515785 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:25:41 crc kubenswrapper[4706]: I1208 19:25:41.564938 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:25:42 crc kubenswrapper[4706]: I1208 19:25:42.334156 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-njntc"] Dec 08 19:25:42 crc kubenswrapper[4706]: I1208 19:25:42.335313 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:25:42 crc kubenswrapper[4706]: I1208 19:25:42.580061 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-njntc" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="registry-server" containerID="cri-o://ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e" gracePeriod=2 Dec 08 19:25:43 crc kubenswrapper[4706]: I1208 19:25:43.204272 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-q6t9t" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="registry-server" probeResult="failure" output=< Dec 08 19:25:43 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:25:43 crc kubenswrapper[4706]: > Dec 08 19:25:43 crc kubenswrapper[4706]: I1208 19:25:43.372388 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-8rl9w" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="registry-server" probeResult="failure" output=< Dec 08 19:25:43 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:25:43 crc kubenswrapper[4706]: > Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.253171 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.301287 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.311107 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.373810 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-catalog-content\") pod \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.373871 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rgbr\" (UniqueName: \"kubernetes.io/projected/0a9bae31-9b10-44fb-b30f-7cfca35cda15-kube-api-access-4rgbr\") pod \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.374065 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-utilities\") pod \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\" (UID: \"0a9bae31-9b10-44fb-b30f-7cfca35cda15\") " Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.375619 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-utilities" (OuterVolumeSpecName: "utilities") pod "0a9bae31-9b10-44fb-b30f-7cfca35cda15" (UID: "0a9bae31-9b10-44fb-b30f-7cfca35cda15"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.385139 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9bae31-9b10-44fb-b30f-7cfca35cda15-kube-api-access-4rgbr" (OuterVolumeSpecName: "kube-api-access-4rgbr") pod "0a9bae31-9b10-44fb-b30f-7cfca35cda15" (UID: "0a9bae31-9b10-44fb-b30f-7cfca35cda15"). InnerVolumeSpecName "kube-api-access-4rgbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.435325 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a9bae31-9b10-44fb-b30f-7cfca35cda15" (UID: "0a9bae31-9b10-44fb-b30f-7cfca35cda15"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.475404 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.475448 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9bae31-9b10-44fb-b30f-7cfca35cda15-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.475466 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rgbr\" (UniqueName: \"kubernetes.io/projected/0a9bae31-9b10-44fb-b30f-7cfca35cda15-kube-api-access-4rgbr\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.592933 4706 generic.go:334] "Generic (PLEG): container finished" podID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerID="ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e" exitCode=0 Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.592989 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-njntc" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.592996 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-njntc" event={"ID":"0a9bae31-9b10-44fb-b30f-7cfca35cda15","Type":"ContainerDied","Data":"ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e"} Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.593127 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-njntc" event={"ID":"0a9bae31-9b10-44fb-b30f-7cfca35cda15","Type":"ContainerDied","Data":"7316c7e08caa2ef92dfff7f512cc163f90b7a01f018653a8386a8dfcf3a1dd41"} Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.593156 4706 scope.go:117] "RemoveContainer" containerID="ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.594941 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.630229 4706 scope.go:117] "RemoveContainer" containerID="25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.636827 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-njntc"] Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.644113 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-njntc"] Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.649739 4706 scope.go:117] "RemoveContainer" containerID="65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.651794 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.669703 4706 scope.go:117] "RemoveContainer" containerID="ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e" Dec 08 19:25:44 crc kubenswrapper[4706]: E1208 19:25:44.670937 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e\": container with ID starting with ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e not found: ID does not exist" containerID="ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.670990 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e"} err="failed to get container status \"ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e\": rpc error: code = NotFound desc = could not find container \"ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e\": container with ID starting with ba688772ca24e43ec1f89cbe3c45d0e6f897f363ad33949a58ae82e427764c2e not found: ID does not exist" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.671022 4706 scope.go:117] "RemoveContainer" containerID="25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee" Dec 08 19:25:44 crc kubenswrapper[4706]: E1208 19:25:44.673734 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee\": container with ID starting with 25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee not found: ID does not exist" containerID="25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.673807 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee"} err="failed to get container status \"25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee\": rpc error: code = NotFound desc = could not find container \"25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee\": container with ID starting with 25661c1245a0911341ef5e8b566a246560c11d7ad1fe0058fabd94bafac614ee not found: ID does not exist" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.674897 4706 scope.go:117] "RemoveContainer" containerID="65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf" Dec 08 19:25:44 crc kubenswrapper[4706]: E1208 19:25:44.682142 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf\": container with ID starting with 65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf not found: ID does not exist" containerID="65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf" Dec 08 19:25:44 crc kubenswrapper[4706]: I1208 19:25:44.682217 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf"} err="failed to get container status \"65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf\": rpc error: code = NotFound desc = could not find container \"65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf\": container with ID starting with 65da4e3519059b7862e781ff95a24aeec373e52a2b46cab6b087d7b1677ba3bf not found: ID does not exist" Dec 08 19:25:45 crc kubenswrapper[4706]: I1208 19:25:45.618430 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" path="/var/lib/kubelet/pods/0a9bae31-9b10-44fb-b30f-7cfca35cda15/volumes" Dec 08 19:25:46 crc kubenswrapper[4706]: I1208 19:25:46.739692 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwksc"] Dec 08 19:25:46 crc kubenswrapper[4706]: I1208 19:25:46.740043 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wwksc" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="registry-server" containerID="cri-o://cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926" gracePeriod=2 Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.152556 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.220138 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsnb5\" (UniqueName: \"kubernetes.io/projected/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-kube-api-access-fsnb5\") pod \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.220205 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-catalog-content\") pod \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.220279 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-utilities\") pod \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\" (UID: \"32aaa66c-bc2d-42e3-9c17-681d58e0ef08\") " Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.221083 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-utilities" (OuterVolumeSpecName: "utilities") pod "32aaa66c-bc2d-42e3-9c17-681d58e0ef08" (UID: "32aaa66c-bc2d-42e3-9c17-681d58e0ef08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.231439 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-kube-api-access-fsnb5" (OuterVolumeSpecName: "kube-api-access-fsnb5") pod "32aaa66c-bc2d-42e3-9c17-681d58e0ef08" (UID: "32aaa66c-bc2d-42e3-9c17-681d58e0ef08"). InnerVolumeSpecName "kube-api-access-fsnb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.322152 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.322195 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsnb5\" (UniqueName: \"kubernetes.io/projected/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-kube-api-access-fsnb5\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.332174 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32aaa66c-bc2d-42e3-9c17-681d58e0ef08" (UID: "32aaa66c-bc2d-42e3-9c17-681d58e0ef08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.423553 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32aaa66c-bc2d-42e3-9c17-681d58e0ef08-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.613802 4706 generic.go:334] "Generic (PLEG): container finished" podID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerID="cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926" exitCode=0 Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.614009 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwksc" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.616126 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwksc" event={"ID":"32aaa66c-bc2d-42e3-9c17-681d58e0ef08","Type":"ContainerDied","Data":"cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926"} Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.616187 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwksc" event={"ID":"32aaa66c-bc2d-42e3-9c17-681d58e0ef08","Type":"ContainerDied","Data":"f3ef61765a27b583a52d7ccb7bed896f0887df696601635f046adce38e2d7ba7"} Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.616215 4706 scope.go:117] "RemoveContainer" containerID="cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.636585 4706 scope.go:117] "RemoveContainer" containerID="6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.660975 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwksc"] Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.663710 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wwksc"] Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.672031 4706 scope.go:117] "RemoveContainer" containerID="374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.685781 4706 scope.go:117] "RemoveContainer" containerID="cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926" Dec 08 19:25:47 crc kubenswrapper[4706]: E1208 19:25:47.686103 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926\": container with ID starting with cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926 not found: ID does not exist" containerID="cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.686140 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926"} err="failed to get container status \"cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926\": rpc error: code = NotFound desc = could not find container \"cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926\": container with ID starting with cecc1fe429484dc345a5f3c6bd1b11d556ec4cd026b05ed190c6b2a97b8f1926 not found: ID does not exist" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.686169 4706 scope.go:117] "RemoveContainer" containerID="6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13" Dec 08 19:25:47 crc kubenswrapper[4706]: E1208 19:25:47.686408 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13\": container with ID starting with 6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13 not found: ID does not exist" containerID="6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.686438 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13"} err="failed to get container status \"6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13\": rpc error: code = NotFound desc = could not find container \"6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13\": container with ID starting with 6c09ad87f770312d7863ea466f2f191e50dbfefe1eeae8136885e068d2a47d13 not found: ID does not exist" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.686454 4706 scope.go:117] "RemoveContainer" containerID="374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d" Dec 08 19:25:47 crc kubenswrapper[4706]: E1208 19:25:47.686710 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d\": container with ID starting with 374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d not found: ID does not exist" containerID="374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d" Dec 08 19:25:47 crc kubenswrapper[4706]: I1208 19:25:47.686743 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d"} err="failed to get container status \"374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d\": rpc error: code = NotFound desc = could not find container \"374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d\": container with ID starting with 374890affd45e0312b424d4f6398390ec492b1d7b195565bdb7d2ea56363d78d not found: ID does not exist" Dec 08 19:25:49 crc kubenswrapper[4706]: I1208 19:25:49.616622 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" path="/var/lib/kubelet/pods/32aaa66c-bc2d-42e3-9c17-681d58e0ef08/volumes" Dec 08 19:25:52 crc kubenswrapper[4706]: I1208 19:25:52.340499 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.010859 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2h6vv"] Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.014078 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2h6vv" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerName="registry-server" containerID="cri-o://09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae" gracePeriod=30 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.018025 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8rl9w"] Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.018255 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8rl9w" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="registry-server" containerID="cri-o://eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c" gracePeriod=30 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.030076 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q6t9t"] Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.030419 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q6t9t" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="registry-server" containerID="cri-o://9b8b6850d6b092bc69c7a6b89e103774a305bfe007226cea7b138b5813292af3" gracePeriod=30 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.049054 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-55zvr"] Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.049304 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" podUID="2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" containerName="marketplace-operator" containerID="cri-o://9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331" gracePeriod=30 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.054606 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fmnd"] Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.054887 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6fmnd" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="registry-server" containerID="cri-o://6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764" gracePeriod=30 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.063495 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zsdd9"] Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.064312 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zsdd9" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="registry-server" containerID="cri-o://fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969" gracePeriod=30 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.076819 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q4hlk"] Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.077077 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="extract-content" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077092 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="extract-content" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.077102 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="registry-server" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077108 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="registry-server" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.077126 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="registry-server" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077132 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="registry-server" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.077141 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="extract-utilities" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077147 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="extract-utilities" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.077156 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="extract-content" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077162 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="extract-content" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.077171 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="extract-utilities" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077179 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="extract-utilities" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077321 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9bae31-9b10-44fb-b30f-7cfca35cda15" containerName="registry-server" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077340 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="32aaa66c-bc2d-42e3-9c17-681d58e0ef08" containerName="registry-server" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.077878 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.090251 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q4hlk"] Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.132775 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p854v\" (UniqueName: \"kubernetes.io/projected/89875d36-6d95-4d46-9fc5-c18d8fccaae3-kube-api-access-p854v\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.132821 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/89875d36-6d95-4d46-9fc5-c18d8fccaae3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.132843 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/89875d36-6d95-4d46-9fc5-c18d8fccaae3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.136713 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8rl9w"] Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.208730 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969 is running failed: container process not found" containerID="fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969" cmd=["grpc_health_probe","-addr=:50051"] Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.209082 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969 is running failed: container process not found" containerID="fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969" cmd=["grpc_health_probe","-addr=:50051"] Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.209744 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969 is running failed: container process not found" containerID="fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969" cmd=["grpc_health_probe","-addr=:50051"] Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.209785 4706 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-zsdd9" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="registry-server" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.234156 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p854v\" (UniqueName: \"kubernetes.io/projected/89875d36-6d95-4d46-9fc5-c18d8fccaae3-kube-api-access-p854v\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.234207 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/89875d36-6d95-4d46-9fc5-c18d8fccaae3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.234238 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/89875d36-6d95-4d46-9fc5-c18d8fccaae3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.236086 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/89875d36-6d95-4d46-9fc5-c18d8fccaae3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.244513 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/89875d36-6d95-4d46-9fc5-c18d8fccaae3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.253388 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p854v\" (UniqueName: \"kubernetes.io/projected/89875d36-6d95-4d46-9fc5-c18d8fccaae3-kube-api-access-p854v\") pod \"marketplace-operator-79b997595-q4hlk\" (UID: \"89875d36-6d95-4d46-9fc5-c18d8fccaae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.545151 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.554199 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.558077 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.575509 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.583111 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.584842 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.641793 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxkgj\" (UniqueName: \"kubernetes.io/projected/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-kube-api-access-mxkgj\") pod \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.641877 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-catalog-content\") pod \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.651116 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-kube-api-access-mxkgj" (OuterVolumeSpecName: "kube-api-access-mxkgj") pod "443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" (UID: "443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5"). InnerVolumeSpecName "kube-api-access-mxkgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.671661 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26f226b0-fa6c-4efc-af99-f69c1ad8a464" (UID: "26f226b0-fa6c-4efc-af99-f69c1ad8a464"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.681753 4706 generic.go:334] "Generic (PLEG): container finished" podID="2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" containerID="9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331" exitCode=0 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.681916 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.681902 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" event={"ID":"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac","Type":"ContainerDied","Data":"9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.681961 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-55zvr" event={"ID":"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac","Type":"ContainerDied","Data":"daf051574b2b7e7c71e14d25cf0108c1c857d59903d3623649f23ca7b725975e"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.681984 4706 scope.go:117] "RemoveContainer" containerID="9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.684678 4706 generic.go:334] "Generic (PLEG): container finished" podID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerID="fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969" exitCode=0 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.684761 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zsdd9" event={"ID":"63da40ba-a150-4396-bb8d-d8e4096a3961","Type":"ContainerDied","Data":"fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.684811 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zsdd9" event={"ID":"63da40ba-a150-4396-bb8d-d8e4096a3961","Type":"ContainerDied","Data":"58ebe40f9802ae8a6088b27eac961fb80c5cf52c35f1fb3be0384df5c8ea7f28"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.685398 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zsdd9" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.700593 4706 generic.go:334] "Generic (PLEG): container finished" podID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerID="6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764" exitCode=0 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.700726 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fmnd" event={"ID":"26f226b0-fa6c-4efc-af99-f69c1ad8a464","Type":"ContainerDied","Data":"6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.700762 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fmnd" event={"ID":"26f226b0-fa6c-4efc-af99-f69c1ad8a464","Type":"ContainerDied","Data":"11cfa80821e6de8c7cf71cce6010fa42d058c9925580148eb63baa5c95750f52"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.700891 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fmnd" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.718946 4706 generic.go:334] "Generic (PLEG): container finished" podID="3200b927-522a-4762-a731-c56780f4a137" containerID="eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c" exitCode=0 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.719305 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8rl9w" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.723922 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rl9w" event={"ID":"3200b927-522a-4762-a731-c56780f4a137","Type":"ContainerDied","Data":"eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.724071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rl9w" event={"ID":"3200b927-522a-4762-a731-c56780f4a137","Type":"ContainerDied","Data":"571f25d3aa36fb8ffe8d314844633b2404d868d9e9d3d89546dc56fcac181167"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.729860 4706 generic.go:334] "Generic (PLEG): container finished" podID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerID="9b8b6850d6b092bc69c7a6b89e103774a305bfe007226cea7b138b5813292af3" exitCode=0 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.729948 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6t9t" event={"ID":"9752bb6b-117b-4e2b-8794-a91e14dbeaf7","Type":"ContainerDied","Data":"9b8b6850d6b092bc69c7a6b89e103774a305bfe007226cea7b138b5813292af3"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.739106 4706 generic.go:334] "Generic (PLEG): container finished" podID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerID="09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae" exitCode=0 Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.739221 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h6vv" event={"ID":"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5","Type":"ContainerDied","Data":"09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.739299 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h6vv" event={"ID":"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5","Type":"ContainerDied","Data":"7a13e5d4c360407b6435bc3d7ff3ef8b0f82d645711859ae9a55f1e839674c62"} Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.739509 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h6vv" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.742114 4706 scope.go:117] "RemoveContainer" containerID="9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.743360 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331\": container with ID starting with 9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331 not found: ID does not exist" containerID="9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.743407 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331"} err="failed to get container status \"9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331\": rpc error: code = NotFound desc = could not find container \"9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331\": container with ID starting with 9ce755c6297cef3e16d1e85900d0ce78411840a2069f1179a5bba75eda98b331 not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.743430 4706 scope.go:117] "RemoveContainer" containerID="fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.743435 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-utilities\") pod \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.743873 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whp92\" (UniqueName: \"kubernetes.io/projected/26f226b0-fa6c-4efc-af99-f69c1ad8a464-kube-api-access-whp92\") pod \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.743952 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qvvg\" (UniqueName: \"kubernetes.io/projected/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-kube-api-access-5qvvg\") pod \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.744137 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5pxf\" (UniqueName: \"kubernetes.io/projected/3200b927-522a-4762-a731-c56780f4a137-kube-api-access-d5pxf\") pod \"3200b927-522a-4762-a731-c56780f4a137\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.744371 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-utilities\") pod \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\" (UID: \"26f226b0-fa6c-4efc-af99-f69c1ad8a464\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.744519 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-operator-metrics\") pod \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.744966 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-utilities\") pod \"3200b927-522a-4762-a731-c56780f4a137\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.745069 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-catalog-content\") pod \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\" (UID: \"443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.745123 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxfrp\" (UniqueName: \"kubernetes.io/projected/63da40ba-a150-4396-bb8d-d8e4096a3961-kube-api-access-lxfrp\") pod \"63da40ba-a150-4396-bb8d-d8e4096a3961\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.745154 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-catalog-content\") pod \"63da40ba-a150-4396-bb8d-d8e4096a3961\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.745308 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-catalog-content\") pod \"3200b927-522a-4762-a731-c56780f4a137\" (UID: \"3200b927-522a-4762-a731-c56780f4a137\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.745384 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-utilities\") pod \"63da40ba-a150-4396-bb8d-d8e4096a3961\" (UID: \"63da40ba-a150-4396-bb8d-d8e4096a3961\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.745551 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-trusted-ca\") pod \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\" (UID: \"2b144ce7-6f82-46a9-ac9d-0a98aa48bbac\") " Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.747026 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxkgj\" (UniqueName: \"kubernetes.io/projected/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-kube-api-access-mxkgj\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.747067 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.748675 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" (UID: "2b144ce7-6f82-46a9-ac9d-0a98aa48bbac"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.749714 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-utilities" (OuterVolumeSpecName: "utilities") pod "26f226b0-fa6c-4efc-af99-f69c1ad8a464" (UID: "26f226b0-fa6c-4efc-af99-f69c1ad8a464"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.751041 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-utilities" (OuterVolumeSpecName: "utilities") pod "443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" (UID: "443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.752075 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-utilities" (OuterVolumeSpecName: "utilities") pod "63da40ba-a150-4396-bb8d-d8e4096a3961" (UID: "63da40ba-a150-4396-bb8d-d8e4096a3961"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.753408 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-utilities" (OuterVolumeSpecName: "utilities") pod "3200b927-522a-4762-a731-c56780f4a137" (UID: "3200b927-522a-4762-a731-c56780f4a137"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.756074 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63da40ba-a150-4396-bb8d-d8e4096a3961-kube-api-access-lxfrp" (OuterVolumeSpecName: "kube-api-access-lxfrp") pod "63da40ba-a150-4396-bb8d-d8e4096a3961" (UID: "63da40ba-a150-4396-bb8d-d8e4096a3961"). InnerVolumeSpecName "kube-api-access-lxfrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.756076 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3200b927-522a-4762-a731-c56780f4a137-kube-api-access-d5pxf" (OuterVolumeSpecName: "kube-api-access-d5pxf") pod "3200b927-522a-4762-a731-c56780f4a137" (UID: "3200b927-522a-4762-a731-c56780f4a137"). InnerVolumeSpecName "kube-api-access-d5pxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.756712 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-kube-api-access-5qvvg" (OuterVolumeSpecName: "kube-api-access-5qvvg") pod "2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" (UID: "2b144ce7-6f82-46a9-ac9d-0a98aa48bbac"). InnerVolumeSpecName "kube-api-access-5qvvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.756768 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f226b0-fa6c-4efc-af99-f69c1ad8a464-kube-api-access-whp92" (OuterVolumeSpecName: "kube-api-access-whp92") pod "26f226b0-fa6c-4efc-af99-f69c1ad8a464" (UID: "26f226b0-fa6c-4efc-af99-f69c1ad8a464"). InnerVolumeSpecName "kube-api-access-whp92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.770821 4706 scope.go:117] "RemoveContainer" containerID="9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.775933 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" (UID: "2b144ce7-6f82-46a9-ac9d-0a98aa48bbac"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.790968 4706 scope.go:117] "RemoveContainer" containerID="aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.814353 4706 scope.go:117] "RemoveContainer" containerID="fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.815227 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969\": container with ID starting with fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969 not found: ID does not exist" containerID="fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.815348 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969"} err="failed to get container status \"fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969\": rpc error: code = NotFound desc = could not find container \"fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969\": container with ID starting with fd5bb27f13c0a72c8d3f65c61c427ebde01821e637024e4f48e4d1b4094b4969 not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.815473 4706 scope.go:117] "RemoveContainer" containerID="9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.816067 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840\": container with ID starting with 9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840 not found: ID does not exist" containerID="9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.816095 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840"} err="failed to get container status \"9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840\": rpc error: code = NotFound desc = could not find container \"9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840\": container with ID starting with 9d7acb0bfe6558664a66caf294ce09bcb74de3e0834154cf28d7439b137df840 not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.816123 4706 scope.go:117] "RemoveContainer" containerID="aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.816599 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347\": container with ID starting with aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347 not found: ID does not exist" containerID="aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.816646 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347"} err="failed to get container status \"aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347\": rpc error: code = NotFound desc = could not find container \"aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347\": container with ID starting with aa5e683a9d4a12a33ce7cb619197fdd49dd3e19cccde90edb2da78aceaa74347 not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.816680 4706 scope.go:117] "RemoveContainer" containerID="6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.832000 4706 scope.go:117] "RemoveContainer" containerID="d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.841805 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" (UID: "443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.848971 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxfrp\" (UniqueName: \"kubernetes.io/projected/63da40ba-a150-4396-bb8d-d8e4096a3961-kube-api-access-lxfrp\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849001 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849016 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849027 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whp92\" (UniqueName: \"kubernetes.io/projected/26f226b0-fa6c-4efc-af99-f69c1ad8a464-kube-api-access-whp92\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849038 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849047 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qvvg\" (UniqueName: \"kubernetes.io/projected/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-kube-api-access-5qvvg\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849056 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5pxf\" (UniqueName: \"kubernetes.io/projected/3200b927-522a-4762-a731-c56780f4a137-kube-api-access-d5pxf\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849065 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f226b0-fa6c-4efc-af99-f69c1ad8a464-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849074 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849084 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.849093 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.858920 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3200b927-522a-4762-a731-c56780f4a137" (UID: "3200b927-522a-4762-a731-c56780f4a137"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.870562 4706 scope.go:117] "RemoveContainer" containerID="2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.890830 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63da40ba-a150-4396-bb8d-d8e4096a3961" (UID: "63da40ba-a150-4396-bb8d-d8e4096a3961"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.895112 4706 scope.go:117] "RemoveContainer" containerID="6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.895774 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764\": container with ID starting with 6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764 not found: ID does not exist" containerID="6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.895826 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764"} err="failed to get container status \"6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764\": rpc error: code = NotFound desc = could not find container \"6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764\": container with ID starting with 6f53ca581f54be956fd11f222c05d44ef4b3d2a00a1e78ed03987e049642f764 not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.895864 4706 scope.go:117] "RemoveContainer" containerID="d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.896731 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494\": container with ID starting with d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494 not found: ID does not exist" containerID="d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.896761 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494"} err="failed to get container status \"d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494\": rpc error: code = NotFound desc = could not find container \"d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494\": container with ID starting with d1d0b0d68618d540370a6e3ef89ff213dbd58ca90d6dfa7de8ff0b35a67d4494 not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.896779 4706 scope.go:117] "RemoveContainer" containerID="2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.897079 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac\": container with ID starting with 2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac not found: ID does not exist" containerID="2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.897116 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac"} err="failed to get container status \"2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac\": rpc error: code = NotFound desc = could not find container \"2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac\": container with ID starting with 2d5a898c8e21bb1d2125a0f531961479311556479c879760209da138993d62ac not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.897132 4706 scope.go:117] "RemoveContainer" containerID="eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.918016 4706 scope.go:117] "RemoveContainer" containerID="f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.937821 4706 scope.go:117] "RemoveContainer" containerID="a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.949807 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3200b927-522a-4762-a731-c56780f4a137-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.949839 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63da40ba-a150-4396-bb8d-d8e4096a3961-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.957614 4706 scope.go:117] "RemoveContainer" containerID="eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.958061 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c\": container with ID starting with eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c not found: ID does not exist" containerID="eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.958094 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c"} err="failed to get container status \"eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c\": rpc error: code = NotFound desc = could not find container \"eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c\": container with ID starting with eb94ac4ad810677e1c228fb55942521a147c30f507f1be666c9f0527e790697c not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.958117 4706 scope.go:117] "RemoveContainer" containerID="f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.958539 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151\": container with ID starting with f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151 not found: ID does not exist" containerID="f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.958565 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151"} err="failed to get container status \"f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151\": rpc error: code = NotFound desc = could not find container \"f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151\": container with ID starting with f52eff3c4320feb1d205bc4afcb21ea80a7ca2adafd24ce6d68edefb239eb151 not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.958581 4706 scope.go:117] "RemoveContainer" containerID="a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b" Dec 08 19:25:54 crc kubenswrapper[4706]: E1208 19:25:54.958975 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b\": container with ID starting with a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b not found: ID does not exist" containerID="a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.958996 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b"} err="failed to get container status \"a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b\": rpc error: code = NotFound desc = could not find container \"a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b\": container with ID starting with a89a5a3dd8322260e6158b50a2aa230c73552f3fddeb30880a1bc245a09e464b not found: ID does not exist" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.959008 4706 scope.go:117] "RemoveContainer" containerID="09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.974357 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:25:54 crc kubenswrapper[4706]: I1208 19:25:54.991169 4706 scope.go:117] "RemoveContainer" containerID="689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.016291 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q4hlk"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.033492 4706 scope.go:117] "RemoveContainer" containerID="cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275" Dec 08 19:25:55 crc kubenswrapper[4706]: W1208 19:25:55.041388 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89875d36_6d95_4d46_9fc5_c18d8fccaae3.slice/crio-a1f758b81e54f357c02a873248896b90ef5b325917c47057cb16a79ef37ced67 WatchSource:0}: Error finding container a1f758b81e54f357c02a873248896b90ef5b325917c47057cb16a79ef37ced67: Status 404 returned error can't find the container with id a1f758b81e54f357c02a873248896b90ef5b325917c47057cb16a79ef37ced67 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.047617 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fmnd"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.049434 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fmnd"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.054715 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-utilities\") pod \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.055095 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w7m4\" (UniqueName: \"kubernetes.io/projected/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-kube-api-access-5w7m4\") pod \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.055129 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-catalog-content\") pod \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\" (UID: \"9752bb6b-117b-4e2b-8794-a91e14dbeaf7\") " Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.063443 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-utilities" (OuterVolumeSpecName: "utilities") pod "9752bb6b-117b-4e2b-8794-a91e14dbeaf7" (UID: "9752bb6b-117b-4e2b-8794-a91e14dbeaf7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.065468 4706 scope.go:117] "RemoveContainer" containerID="09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.065471 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-kube-api-access-5w7m4" (OuterVolumeSpecName: "kube-api-access-5w7m4") pod "9752bb6b-117b-4e2b-8794-a91e14dbeaf7" (UID: "9752bb6b-117b-4e2b-8794-a91e14dbeaf7"). InnerVolumeSpecName "kube-api-access-5w7m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.066218 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae\": container with ID starting with 09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae not found: ID does not exist" containerID="09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.066272 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae"} err="failed to get container status \"09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae\": rpc error: code = NotFound desc = could not find container \"09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae\": container with ID starting with 09140bfd32466e296cd1988cfd948f43b635a25a49aa38cc9067839787f6a1ae not found: ID does not exist" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.066302 4706 scope.go:117] "RemoveContainer" containerID="689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.073471 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zsdd9"] Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.076882 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c\": container with ID starting with 689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c not found: ID does not exist" containerID="689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.076926 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c"} err="failed to get container status \"689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c\": rpc error: code = NotFound desc = could not find container \"689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c\": container with ID starting with 689a8e400e7050df869d647235d1c10e203b562809d9c72e5d3d08f193999d4c not found: ID does not exist" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.076961 4706 scope.go:117] "RemoveContainer" containerID="cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.078567 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275\": container with ID starting with cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275 not found: ID does not exist" containerID="cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.078619 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275"} err="failed to get container status \"cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275\": rpc error: code = NotFound desc = could not find container \"cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275\": container with ID starting with cd7a586883a3073745115d051f93ad19304f3acab997cf1b99239d1e743ca275 not found: ID does not exist" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.079114 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zsdd9"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.081520 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-55zvr"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.083951 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-55zvr"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.107896 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8rl9w"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.111024 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8rl9w"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.116186 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2h6vv"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.123438 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2h6vv"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.154772 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9752bb6b-117b-4e2b-8794-a91e14dbeaf7" (UID: "9752bb6b-117b-4e2b-8794-a91e14dbeaf7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.157521 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w7m4\" (UniqueName: \"kubernetes.io/projected/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-kube-api-access-5w7m4\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.157576 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.157590 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9752bb6b-117b-4e2b-8794-a91e14dbeaf7-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.335579 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336389 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336413 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336427 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336438 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336447 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336455 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336465 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" containerName="marketplace-operator" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336472 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" containerName="marketplace-operator" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336483 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336490 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336500 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336509 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336520 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336529 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336539 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336548 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336561 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336569 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336582 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336589 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336598 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336605 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336620 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336630 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="extract-content" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336640 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336647 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336658 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336665 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336676 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336684 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="extract-utilities" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.336693 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336700 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336797 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" containerName="marketplace-operator" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.336812 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3200b927-522a-4762-a731-c56780f4a137" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.337098 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.337110 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.337118 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.337129 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" containerName="registry-server" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.337562 4706 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.337584 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.337676 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338012 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1" gracePeriod=15 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338068 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996" gracePeriod=15 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338169 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8" gracePeriod=15 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338160 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1" gracePeriod=15 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338237 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac" gracePeriod=15 Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.338357 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338832 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.338848 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338856 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.338864 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338870 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.338881 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338889 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.338897 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338906 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.338915 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338920 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.338927 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.338933 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.339019 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.339029 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.339036 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.339044 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.339051 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.339060 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.341287 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.360443 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.360883 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.360962 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.361045 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.361116 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.361207 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.361325 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.361433 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.370753 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.462756 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.462812 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.462835 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.462857 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.462887 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.462915 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.462935 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.462981 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.463048 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.463091 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.463116 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.463139 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.463159 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.463189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.463213 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.463236 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.614764 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f226b0-fa6c-4efc-af99-f69c1ad8a464" path="/var/lib/kubelet/pods/26f226b0-fa6c-4efc-af99-f69c1ad8a464/volumes" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.615584 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b144ce7-6f82-46a9-ac9d-0a98aa48bbac" path="/var/lib/kubelet/pods/2b144ce7-6f82-46a9-ac9d-0a98aa48bbac/volumes" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.616059 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3200b927-522a-4762-a731-c56780f4a137" path="/var/lib/kubelet/pods/3200b927-522a-4762-a731-c56780f4a137/volumes" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.616781 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5" path="/var/lib/kubelet/pods/443b0c73-a6b1-4f48-b46c-7eae5ff8c7c5/volumes" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.617381 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63da40ba-a150-4396-bb8d-d8e4096a3961" path="/var/lib/kubelet/pods/63da40ba-a150-4396-bb8d-d8e4096a3961/volumes" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.664946 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:25:55 crc kubenswrapper[4706]: W1208 19:25:55.735547 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-b7d032194b4acdd51fe416b6b34ac56fe38252ad34588abf70cbed8810fd33f4 WatchSource:0}: Error finding container b7d032194b4acdd51fe416b6b34ac56fe38252ad34588abf70cbed8810fd33f4: Status 404 returned error can't find the container with id b7d032194b4acdd51fe416b6b34ac56fe38252ad34588abf70cbed8810fd33f4 Dec 08 19:25:55 crc kubenswrapper[4706]: E1208 19:25:55.738949 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f54042580325f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-08 19:25:55.738284639 +0000 UTC m=+238.380485642,LastTimestamp:2025-12-08 19:25:55.738284639 +0000 UTC m=+238.380485642,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.748910 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/0.log" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.749314 4706 generic.go:334] "Generic (PLEG): container finished" podID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" containerID="bce8768f1558e3ce8893d42c4c4f38be6c88069a603b90e85070488228b28329" exitCode=1 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.749654 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" event={"ID":"89875d36-6d95-4d46-9fc5-c18d8fccaae3","Type":"ContainerDied","Data":"bce8768f1558e3ce8893d42c4c4f38be6c88069a603b90e85070488228b28329"} Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.749698 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" event={"ID":"89875d36-6d95-4d46-9fc5-c18d8fccaae3","Type":"ContainerStarted","Data":"a1f758b81e54f357c02a873248896b90ef5b325917c47057cb16a79ef37ced67"} Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.749890 4706 scope.go:117] "RemoveContainer" containerID="bce8768f1558e3ce8893d42c4c4f38be6c88069a603b90e85070488228b28329" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.751129 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.751639 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.755515 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6t9t" event={"ID":"9752bb6b-117b-4e2b-8794-a91e14dbeaf7","Type":"ContainerDied","Data":"c6ff3ca1d15d49c553a618a7b2cc37e5a9787ba282073734b58cf041d88639f1"} Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.755626 4706 scope.go:117] "RemoveContainer" containerID="9b8b6850d6b092bc69c7a6b89e103774a305bfe007226cea7b138b5813292af3" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.755813 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q6t9t" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.757259 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.757577 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.757860 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.758992 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.759301 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.759601 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.764349 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.770138 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.770943 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1" exitCode=0 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.771072 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996" exitCode=0 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.771174 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac" exitCode=0 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.771297 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8" exitCode=2 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.772619 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b7d032194b4acdd51fe416b6b34ac56fe38252ad34588abf70cbed8810fd33f4"} Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.780820 4706 generic.go:334] "Generic (PLEG): container finished" podID="1d87b9a0-63a8-4619-88cb-127f2608edd6" containerID="60df5744296e422b2735860df0b42e8d7235bac12190d4f93629d5ea63a13a68" exitCode=0 Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.780919 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1d87b9a0-63a8-4619-88cb-127f2608edd6","Type":"ContainerDied","Data":"60df5744296e422b2735860df0b42e8d7235bac12190d4f93629d5ea63a13a68"} Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.782495 4706 scope.go:117] "RemoveContainer" containerID="06466ad7b3b1602833841a787d961f883ae924716ae11da2976838092631ea05" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.784175 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.785902 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.786235 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.786583 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.800108 4706 scope.go:117] "RemoveContainer" containerID="1554d09d7a70534828b57a5c1ab61b370ecaf946344e71188cac78cff8e3403f" Dec 08 19:25:55 crc kubenswrapper[4706]: I1208 19:25:55.819131 4706 scope.go:117] "RemoveContainer" containerID="269346f43e1c7d406de7d8e05940adf8db62105e9d227165f5fff8ed0fa2aaa0" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.800806 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.803307 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"71a60de3f04c9a2613bd658290cb3f010b3e0c44d5c2fc77fe444660a8c05d12"} Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.805241 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.805453 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.805662 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.805840 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.807336 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/1.log" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.807869 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/0.log" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.807923 4706 generic.go:334] "Generic (PLEG): container finished" podID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" containerID="cabfd9cb24d1854ff19ac23c1a8f7be9cb01e85bb4d646a2d611c7917e1d8bdd" exitCode=1 Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.807994 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" event={"ID":"89875d36-6d95-4d46-9fc5-c18d8fccaae3","Type":"ContainerDied","Data":"cabfd9cb24d1854ff19ac23c1a8f7be9cb01e85bb4d646a2d611c7917e1d8bdd"} Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.808040 4706 scope.go:117] "RemoveContainer" containerID="bce8768f1558e3ce8893d42c4c4f38be6c88069a603b90e85070488228b28329" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.808680 4706 scope.go:117] "RemoveContainer" containerID="cabfd9cb24d1854ff19ac23c1a8f7be9cb01e85bb4d646a2d611c7917e1d8bdd" Dec 08 19:25:56 crc kubenswrapper[4706]: E1208 19:25:56.808899 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-q4hlk_openshift-marketplace(89875d36-6d95-4d46-9fc5-c18d8fccaae3)\"" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.809663 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.810663 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.810994 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:56 crc kubenswrapper[4706]: I1208 19:25:56.811238 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.109584 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.110706 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.110983 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.111253 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.111525 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.189301 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d87b9a0-63a8-4619-88cb-127f2608edd6-kube-api-access\") pod \"1d87b9a0-63a8-4619-88cb-127f2608edd6\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.189400 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-var-lock\") pod \"1d87b9a0-63a8-4619-88cb-127f2608edd6\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.189435 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-kubelet-dir\") pod \"1d87b9a0-63a8-4619-88cb-127f2608edd6\" (UID: \"1d87b9a0-63a8-4619-88cb-127f2608edd6\") " Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.189536 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-var-lock" (OuterVolumeSpecName: "var-lock") pod "1d87b9a0-63a8-4619-88cb-127f2608edd6" (UID: "1d87b9a0-63a8-4619-88cb-127f2608edd6"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.189622 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1d87b9a0-63a8-4619-88cb-127f2608edd6" (UID: "1d87b9a0-63a8-4619-88cb-127f2608edd6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.189874 4706 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-var-lock\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.189904 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d87b9a0-63a8-4619-88cb-127f2608edd6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.194800 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d87b9a0-63a8-4619-88cb-127f2608edd6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1d87b9a0-63a8-4619-88cb-127f2608edd6" (UID: "1d87b9a0-63a8-4619-88cb-127f2608edd6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.291446 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d87b9a0-63a8-4619-88cb-127f2608edd6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.610898 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.612888 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.613305 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.613532 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.816833 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.818497 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.820117 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.820493 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.821009 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.821283 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.821353 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.822055 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1" exitCode=0 Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.822064 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.822184 4706 scope.go:117] "RemoveContainer" containerID="ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.824587 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/1.log" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.825530 4706 scope.go:117] "RemoveContainer" containerID="cabfd9cb24d1854ff19ac23c1a8f7be9cb01e85bb4d646a2d611c7917e1d8bdd" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.825697 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: E1208 19:25:57.825983 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-q4hlk_openshift-marketplace(89875d36-6d95-4d46-9fc5-c18d8fccaae3)\"" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.826183 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.826577 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1d87b9a0-63a8-4619-88cb-127f2608edd6","Type":"ContainerDied","Data":"768d2e67ef4da395b788b1b59293723eb42bef1bd920542cd84f2d0e6b788624"} Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.826679 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="768d2e67ef4da395b788b1b59293723eb42bef1bd920542cd84f2d0e6b788624" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.826704 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.826596 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.827048 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.827359 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.832156 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.833131 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.833483 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.833856 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.834082 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.842213 4706 scope.go:117] "RemoveContainer" containerID="292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.858928 4706 scope.go:117] "RemoveContainer" containerID="e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.875313 4706 scope.go:117] "RemoveContainer" containerID="641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.890537 4706 scope.go:117] "RemoveContainer" containerID="1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900180 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900315 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900395 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900474 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900496 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900538 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900828 4706 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900852 4706 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.900863 4706 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.906756 4706 scope.go:117] "RemoveContainer" containerID="968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.923839 4706 scope.go:117] "RemoveContainer" containerID="ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1" Dec 08 19:25:57 crc kubenswrapper[4706]: E1208 19:25:57.924671 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\": container with ID starting with ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1 not found: ID does not exist" containerID="ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.924720 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1"} err="failed to get container status \"ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\": rpc error: code = NotFound desc = could not find container \"ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1\": container with ID starting with ac764a9f126997db595a79fba49f3016855688f596f1fbe549d69d6e60d5aff1 not found: ID does not exist" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.924773 4706 scope.go:117] "RemoveContainer" containerID="292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996" Dec 08 19:25:57 crc kubenswrapper[4706]: E1208 19:25:57.925088 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\": container with ID starting with 292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996 not found: ID does not exist" containerID="292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.925131 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996"} err="failed to get container status \"292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\": rpc error: code = NotFound desc = could not find container \"292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996\": container with ID starting with 292f23bd1739455dd550b2fb8380d220def10683120540c18004d3ebd40a4996 not found: ID does not exist" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.925150 4706 scope.go:117] "RemoveContainer" containerID="e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac" Dec 08 19:25:57 crc kubenswrapper[4706]: E1208 19:25:57.925490 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\": container with ID starting with e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac not found: ID does not exist" containerID="e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.925523 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac"} err="failed to get container status \"e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\": rpc error: code = NotFound desc = could not find container \"e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac\": container with ID starting with e6aefdee40ec9edd9466548239dd84f90a6ea4bf56ee7cc3ab1a3a8bda39a7ac not found: ID does not exist" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.925543 4706 scope.go:117] "RemoveContainer" containerID="641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8" Dec 08 19:25:57 crc kubenswrapper[4706]: E1208 19:25:57.925815 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\": container with ID starting with 641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8 not found: ID does not exist" containerID="641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.925854 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8"} err="failed to get container status \"641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\": rpc error: code = NotFound desc = could not find container \"641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8\": container with ID starting with 641fe6c93904b482831d92b45bb4eeb5ad9787e9161b8023f5554b96befe01f8 not found: ID does not exist" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.925874 4706 scope.go:117] "RemoveContainer" containerID="1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1" Dec 08 19:25:57 crc kubenswrapper[4706]: E1208 19:25:57.926141 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\": container with ID starting with 1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1 not found: ID does not exist" containerID="1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.926171 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1"} err="failed to get container status \"1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\": rpc error: code = NotFound desc = could not find container \"1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1\": container with ID starting with 1f81f1c88ad3a2e33b3510f775484fd837a6aac3aa90617e3d0f942d70605bb1 not found: ID does not exist" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.926189 4706 scope.go:117] "RemoveContainer" containerID="968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb" Dec 08 19:25:57 crc kubenswrapper[4706]: E1208 19:25:57.927325 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\": container with ID starting with 968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb not found: ID does not exist" containerID="968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb" Dec 08 19:25:57 crc kubenswrapper[4706]: I1208 19:25:57.927421 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb"} err="failed to get container status \"968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\": rpc error: code = NotFound desc = could not find container \"968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb\": container with ID starting with 968397ac37474dc42cb944ec137f2a0ae6898e8dd0ef8bf397c62b78078578cb not found: ID does not exist" Dec 08 19:25:58 crc kubenswrapper[4706]: I1208 19:25:58.849048 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:25:58 crc kubenswrapper[4706]: I1208 19:25:58.876524 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:58 crc kubenswrapper[4706]: I1208 19:25:58.877110 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:58 crc kubenswrapper[4706]: I1208 19:25:58.877355 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:58 crc kubenswrapper[4706]: I1208 19:25:58.885849 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:58 crc kubenswrapper[4706]: I1208 19:25:58.886235 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:25:59 crc kubenswrapper[4706]: I1208 19:25:59.621015 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 08 19:26:01 crc kubenswrapper[4706]: E1208 19:26:01.427464 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:01 crc kubenswrapper[4706]: E1208 19:26:01.428592 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:01 crc kubenswrapper[4706]: E1208 19:26:01.429573 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:01 crc kubenswrapper[4706]: E1208 19:26:01.430463 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:01 crc kubenswrapper[4706]: E1208 19:26:01.430977 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:01 crc kubenswrapper[4706]: I1208 19:26:01.431023 4706 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 08 19:26:01 crc kubenswrapper[4706]: E1208 19:26:01.431392 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="200ms" Dec 08 19:26:01 crc kubenswrapper[4706]: E1208 19:26:01.632764 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="400ms" Dec 08 19:26:02 crc kubenswrapper[4706]: E1208 19:26:02.034076 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="800ms" Dec 08 19:26:02 crc kubenswrapper[4706]: E1208 19:26:02.835449 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="1.6s" Dec 08 19:26:04 crc kubenswrapper[4706]: E1208 19:26:04.187903 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f54042580325f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-08 19:25:55.738284639 +0000 UTC m=+238.380485642,LastTimestamp:2025-12-08 19:25:55.738284639 +0000 UTC m=+238.380485642,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 08 19:26:04 crc kubenswrapper[4706]: E1208 19:26:04.436328 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="3.2s" Dec 08 19:26:04 crc kubenswrapper[4706]: I1208 19:26:04.546810 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:26:04 crc kubenswrapper[4706]: I1208 19:26:04.546929 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:26:04 crc kubenswrapper[4706]: I1208 19:26:04.548136 4706 scope.go:117] "RemoveContainer" containerID="cabfd9cb24d1854ff19ac23c1a8f7be9cb01e85bb4d646a2d611c7917e1d8bdd" Dec 08 19:26:04 crc kubenswrapper[4706]: E1208 19:26:04.548983 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-q4hlk_openshift-marketplace(89875d36-6d95-4d46-9fc5-c18d8fccaae3)\"" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" Dec 08 19:26:07 crc kubenswrapper[4706]: I1208 19:26:07.610299 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:07 crc kubenswrapper[4706]: I1208 19:26:07.611000 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:07 crc kubenswrapper[4706]: I1208 19:26:07.611216 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:07 crc kubenswrapper[4706]: I1208 19:26:07.611558 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:07 crc kubenswrapper[4706]: E1208 19:26:07.637309 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="6.4s" Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.911716 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.912174 4706 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114" exitCode=1 Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.912228 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114"} Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.913090 4706 scope.go:117] "RemoveContainer" containerID="db857715abb8a6b36b55a478a789245d7d7391e2292f02d81f797d3070b26114" Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.914586 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.915116 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.915829 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.916625 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:08 crc kubenswrapper[4706]: I1208 19:26:08.917086 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:09 crc kubenswrapper[4706]: I1208 19:26:09.924820 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 08 19:26:09 crc kubenswrapper[4706]: I1208 19:26:09.925322 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7749dbad2af9f6069413d28cc0d1aa76645227e368a2c514c7d91205c37bdd45"} Dec 08 19:26:09 crc kubenswrapper[4706]: I1208 19:26:09.927783 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:09 crc kubenswrapper[4706]: I1208 19:26:09.928606 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:09 crc kubenswrapper[4706]: I1208 19:26:09.929384 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:09 crc kubenswrapper[4706]: I1208 19:26:09.929966 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:09 crc kubenswrapper[4706]: I1208 19:26:09.930645 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.608184 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.609457 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.609769 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.610052 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.610318 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.610511 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.622026 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.622051 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:10 crc kubenswrapper[4706]: E1208 19:26:10.622707 4706 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.623896 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:10 crc kubenswrapper[4706]: W1208 19:26:10.641861 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-92022bd702c4532bc47abe7b644efcf32454fdf8dde6a04e726a608a9204f093 WatchSource:0}: Error finding container 92022bd702c4532bc47abe7b644efcf32454fdf8dde6a04e726a608a9204f093: Status 404 returned error can't find the container with id 92022bd702c4532bc47abe7b644efcf32454fdf8dde6a04e726a608a9204f093 Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.932142 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"261348d9664f9647b9328f23d669171771ec9c020bd1fe5becdb54e05fccfef3"} Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.932597 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"92022bd702c4532bc47abe7b644efcf32454fdf8dde6a04e726a608a9204f093"} Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.932899 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.932916 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:10 crc kubenswrapper[4706]: E1208 19:26:10.935165 4706 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.935456 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.936102 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.936746 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.937612 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:10 crc kubenswrapper[4706]: I1208 19:26:10.938116 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: E1208 19:26:11.688712 4706 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.113:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" volumeName="registry-storage" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.815110 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.820335 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.821199 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.821816 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.822490 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.823275 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.823708 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.950922 4706 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="261348d9664f9647b9328f23d669171771ec9c020bd1fe5becdb54e05fccfef3" exitCode=0 Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.951071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"261348d9664f9647b9328f23d669171771ec9c020bd1fe5becdb54e05fccfef3"} Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.951600 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.952053 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.952098 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.952888 4706 status_manager.go:851] "Failed to get status for pod" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-q4hlk\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: E1208 19:26:11.952894 4706 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.953765 4706 status_manager.go:851] "Failed to get status for pod" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" pod="openshift-marketplace/community-operators-q6t9t" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-q6t9t\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.954327 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.954810 4706 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:11 crc kubenswrapper[4706]: I1208 19:26:11.955392 4706 status_manager.go:851] "Failed to get status for pod" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Dec 08 19:26:12 crc kubenswrapper[4706]: I1208 19:26:12.960028 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e9b605c407877187598c8a6f1ee949643d73a2ba4d2e2b78650389117159ec0c"} Dec 08 19:26:12 crc kubenswrapper[4706]: I1208 19:26:12.960419 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ff0d5041bafb52e45f7bece169dd5ce486ac894db2cee074be1deb167b33862f"} Dec 08 19:26:12 crc kubenswrapper[4706]: I1208 19:26:12.960431 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"621df82c5ff599aa862e722d982123f8314956197863e34d580e2c924bc330f6"} Dec 08 19:26:12 crc kubenswrapper[4706]: I1208 19:26:12.960439 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dd479795c361cebadd534d0a00833d4a93888fc6d7dfd3594a6f6a83b3c95214"} Dec 08 19:26:13 crc kubenswrapper[4706]: I1208 19:26:13.969847 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9ba2cf3ab99e462b8eded2dedd8ebc9195fc87d73f6dff19f492b8c07e22b5ee"} Dec 08 19:26:13 crc kubenswrapper[4706]: I1208 19:26:13.970283 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:13 crc kubenswrapper[4706]: I1208 19:26:13.970425 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:13 crc kubenswrapper[4706]: I1208 19:26:13.971441 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:14 crc kubenswrapper[4706]: I1208 19:26:14.609102 4706 scope.go:117] "RemoveContainer" containerID="cabfd9cb24d1854ff19ac23c1a8f7be9cb01e85bb4d646a2d611c7917e1d8bdd" Dec 08 19:26:14 crc kubenswrapper[4706]: I1208 19:26:14.977469 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/1.log" Dec 08 19:26:14 crc kubenswrapper[4706]: I1208 19:26:14.978085 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" event={"ID":"89875d36-6d95-4d46-9fc5-c18d8fccaae3","Type":"ContainerStarted","Data":"04705c5938fc3805cdde7360989dd5cad4e55df5893d9868948af3dfb8054ca9"} Dec 08 19:26:14 crc kubenswrapper[4706]: I1208 19:26:14.978449 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:26:14 crc kubenswrapper[4706]: I1208 19:26:14.980195 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q4hlk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Dec 08 19:26:14 crc kubenswrapper[4706]: I1208 19:26:14.980297 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Dec 08 19:26:15 crc kubenswrapper[4706]: I1208 19:26:15.624956 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:15 crc kubenswrapper[4706]: I1208 19:26:15.625022 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:15 crc kubenswrapper[4706]: I1208 19:26:15.633967 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:16 crc kubenswrapper[4706]: I1208 19:26:16.024863 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/2.log" Dec 08 19:26:16 crc kubenswrapper[4706]: I1208 19:26:16.026163 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/1.log" Dec 08 19:26:16 crc kubenswrapper[4706]: I1208 19:26:16.026233 4706 generic.go:334] "Generic (PLEG): container finished" podID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" containerID="04705c5938fc3805cdde7360989dd5cad4e55df5893d9868948af3dfb8054ca9" exitCode=1 Dec 08 19:26:16 crc kubenswrapper[4706]: I1208 19:26:16.026292 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" event={"ID":"89875d36-6d95-4d46-9fc5-c18d8fccaae3","Type":"ContainerDied","Data":"04705c5938fc3805cdde7360989dd5cad4e55df5893d9868948af3dfb8054ca9"} Dec 08 19:26:16 crc kubenswrapper[4706]: I1208 19:26:16.026339 4706 scope.go:117] "RemoveContainer" containerID="cabfd9cb24d1854ff19ac23c1a8f7be9cb01e85bb4d646a2d611c7917e1d8bdd" Dec 08 19:26:16 crc kubenswrapper[4706]: I1208 19:26:16.028848 4706 scope.go:117] "RemoveContainer" containerID="04705c5938fc3805cdde7360989dd5cad4e55df5893d9868948af3dfb8054ca9" Dec 08 19:26:16 crc kubenswrapper[4706]: E1208 19:26:16.029156 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-q4hlk_openshift-marketplace(89875d36-6d95-4d46-9fc5-c18d8fccaae3)\"" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" Dec 08 19:26:17 crc kubenswrapper[4706]: I1208 19:26:17.033871 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/2.log" Dec 08 19:26:17 crc kubenswrapper[4706]: I1208 19:26:17.034497 4706 scope.go:117] "RemoveContainer" containerID="04705c5938fc3805cdde7360989dd5cad4e55df5893d9868948af3dfb8054ca9" Dec 08 19:26:17 crc kubenswrapper[4706]: E1208 19:26:17.034816 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-q4hlk_openshift-marketplace(89875d36-6d95-4d46-9fc5-c18d8fccaae3)\"" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" Dec 08 19:26:19 crc kubenswrapper[4706]: I1208 19:26:19.016882 4706 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:19 crc kubenswrapper[4706]: I1208 19:26:19.050927 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:19 crc kubenswrapper[4706]: I1208 19:26:19.050972 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:19 crc kubenswrapper[4706]: I1208 19:26:19.055171 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:19 crc kubenswrapper[4706]: I1208 19:26:19.169994 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="c0600047-5d54-4c78-a664-382846471149" Dec 08 19:26:20 crc kubenswrapper[4706]: I1208 19:26:20.056307 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:20 crc kubenswrapper[4706]: I1208 19:26:20.057256 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:20 crc kubenswrapper[4706]: I1208 19:26:20.059961 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="c0600047-5d54-4c78-a664-382846471149" Dec 08 19:26:24 crc kubenswrapper[4706]: I1208 19:26:24.545825 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:26:24 crc kubenswrapper[4706]: I1208 19:26:24.547358 4706 scope.go:117] "RemoveContainer" containerID="04705c5938fc3805cdde7360989dd5cad4e55df5893d9868948af3dfb8054ca9" Dec 08 19:26:24 crc kubenswrapper[4706]: E1208 19:26:24.547708 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-q4hlk_openshift-marketplace(89875d36-6d95-4d46-9fc5-c18d8fccaae3)\"" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" Dec 08 19:26:25 crc kubenswrapper[4706]: I1208 19:26:25.768042 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 19:26:28 crc kubenswrapper[4706]: I1208 19:26:28.928408 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 08 19:26:29 crc kubenswrapper[4706]: I1208 19:26:29.906158 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 08 19:26:30 crc kubenswrapper[4706]: I1208 19:26:30.174436 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 08 19:26:30 crc kubenswrapper[4706]: I1208 19:26:30.423141 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 08 19:26:30 crc kubenswrapper[4706]: I1208 19:26:30.478545 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 08 19:26:31 crc kubenswrapper[4706]: I1208 19:26:31.114138 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 08 19:26:31 crc kubenswrapper[4706]: I1208 19:26:31.166530 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 08 19:26:31 crc kubenswrapper[4706]: I1208 19:26:31.447302 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 08 19:26:31 crc kubenswrapper[4706]: I1208 19:26:31.492948 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 08 19:26:31 crc kubenswrapper[4706]: I1208 19:26:31.756169 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 08 19:26:31 crc kubenswrapper[4706]: I1208 19:26:31.797646 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.050028 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.055733 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.059021 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.102612 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.126119 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.144836 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.210255 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.219276 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.269005 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.304557 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.498723 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.559841 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.672719 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.677616 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.728113 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.823057 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.835884 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.885742 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.937184 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 08 19:26:32 crc kubenswrapper[4706]: I1208 19:26:32.956822 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.015808 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.041954 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.093349 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.107957 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.211828 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.320798 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.325353 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.364541 4706 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.388507 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.399610 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.443898 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.543334 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.547837 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.596881 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.745993 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.773573 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.797502 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.808105 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.835013 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.862606 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 08 19:26:33 crc kubenswrapper[4706]: I1208 19:26:33.960079 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.011825 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.013758 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.019155 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.092893 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.143050 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.179798 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.226483 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.318162 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.418249 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.421036 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.490882 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.537996 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.668435 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.769713 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.777840 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.793754 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.796536 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.825981 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.901148 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 08 19:26:34 crc kubenswrapper[4706]: I1208 19:26:34.962570 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.076688 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.161612 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.275406 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.279630 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.329301 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.348579 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.491526 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.608536 4706 scope.go:117] "RemoveContainer" containerID="04705c5938fc3805cdde7360989dd5cad4e55df5893d9868948af3dfb8054ca9" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.713584 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.734013 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.818221 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.987419 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 08 19:26:35 crc kubenswrapper[4706]: I1208 19:26:35.996492 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.079767 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.150424 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.153292 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/2.log" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.153368 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" event={"ID":"89875d36-6d95-4d46-9fc5-c18d8fccaae3","Type":"ContainerStarted","Data":"945f6ab12830d62804881986e1c83d22e629ad70bc55b40bb4cfea9acea71f12"} Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.153897 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.155942 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q4hlk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.155986 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podUID="89875d36-6d95-4d46-9fc5-c18d8fccaae3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.355436 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.374807 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.466067 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.486509 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.644131 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.764975 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.834883 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.849379 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.870070 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.905822 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.985536 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 08 19:26:36 crc kubenswrapper[4706]: I1208 19:26:36.997303 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.133963 4706 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.164962 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.201717 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.248064 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.251072 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.253189 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.319818 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.323147 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.361845 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.414599 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.536966 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.551571 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.566371 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.624127 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.688117 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.745784 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.851600 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.883103 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.886474 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 08 19:26:37 crc kubenswrapper[4706]: I1208 19:26:37.981625 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.005902 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.054108 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.101065 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.135022 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.139233 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.175510 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.203600 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.442311 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.490657 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.609166 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.657071 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.679041 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.706406 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.804879 4706 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.808718 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=43.808614371 podStartE2EDuration="43.808614371s" podCreationTimestamp="2025-12-08 19:25:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:26:19.108866269 +0000 UTC m=+261.751067272" watchObservedRunningTime="2025-12-08 19:26:38.808614371 +0000 UTC m=+281.450815384" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.813795 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-q4hlk" podStartSLOduration=44.813768875 podStartE2EDuration="44.813768875s" podCreationTimestamp="2025-12-08 19:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:26:36.169193676 +0000 UTC m=+278.811394689" watchObservedRunningTime="2025-12-08 19:26:38.813768875 +0000 UTC m=+281.455969878" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.814771 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/community-operators-q6t9t"] Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.815219 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.815394 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.815426 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="504b4535-ae3b-440f-acc3-2729c2f9713e" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.822466 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.828582 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.828664 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.871054 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.871020184 podStartE2EDuration="19.871020184s" podCreationTimestamp="2025-12-08 19:26:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:26:38.866923074 +0000 UTC m=+281.509124097" watchObservedRunningTime="2025-12-08 19:26:38.871020184 +0000 UTC m=+281.513221187" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.874702 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.895029 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 08 19:26:38 crc kubenswrapper[4706]: I1208 19:26:38.964390 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.120187 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.208910 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.297148 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.298186 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.322013 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.399179 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.470287 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.480480 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.615359 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9752bb6b-117b-4e2b-8794-a91e14dbeaf7" path="/var/lib/kubelet/pods/9752bb6b-117b-4e2b-8794-a91e14dbeaf7/volumes" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.698089 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.699224 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.725427 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.750818 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.769217 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.800185 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.808888 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.811903 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.813664 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.880242 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 08 19:26:39 crc kubenswrapper[4706]: I1208 19:26:39.932964 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.002748 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.023581 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.139647 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.142653 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.285587 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.337614 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.357981 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.393811 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.481801 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.515645 4706 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.516147 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://71a60de3f04c9a2613bd658290cb3f010b3e0c44d5c2fc77fe444660a8c05d12" gracePeriod=5 Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.548469 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.658957 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.659340 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.698336 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.765372 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.807912 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.823482 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.831286 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 08 19:26:40 crc kubenswrapper[4706]: I1208 19:26:40.834597 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.049870 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.064348 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.091040 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.152101 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.157004 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.190504 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.213208 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.214726 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.220397 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.236394 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.257012 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.270838 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.284047 4706 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.467964 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.485779 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.577128 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.597065 4706 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.617210 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.618332 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.635388 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.674643 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.684311 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.807934 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.836345 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.892654 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.895195 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.902741 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.908194 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 08 19:26:41 crc kubenswrapper[4706]: I1208 19:26:41.981626 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.103536 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.192098 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.208647 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.303976 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.321651 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.326304 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.357059 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.419548 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.484589 4706 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.488375 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.532050 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.625068 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.667562 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.683781 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.833686 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.946239 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 08 19:26:42 crc kubenswrapper[4706]: I1208 19:26:42.964517 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.000808 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.016776 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.024897 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.354806 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.468990 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.520172 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.625827 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.630099 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 08 19:26:43 crc kubenswrapper[4706]: I1208 19:26:43.765234 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 08 19:26:44 crc kubenswrapper[4706]: I1208 19:26:44.030959 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 08 19:26:44 crc kubenswrapper[4706]: I1208 19:26:44.297922 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 08 19:26:44 crc kubenswrapper[4706]: I1208 19:26:44.705705 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 08 19:26:44 crc kubenswrapper[4706]: I1208 19:26:44.945102 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 08 19:26:46 crc kubenswrapper[4706]: I1208 19:26:46.210801 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 08 19:26:46 crc kubenswrapper[4706]: I1208 19:26:46.211184 4706 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="71a60de3f04c9a2613bd658290cb3f010b3e0c44d5c2fc77fe444660a8c05d12" exitCode=137 Dec 08 19:26:46 crc kubenswrapper[4706]: I1208 19:26:46.211231 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7d032194b4acdd51fe416b6b34ac56fe38252ad34588abf70cbed8810fd33f4" Dec 08 19:26:46 crc kubenswrapper[4706]: I1208 19:26:46.944776 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 08 19:26:46 crc kubenswrapper[4706]: I1208 19:26:46.945242 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050064 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050110 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050179 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050205 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050195 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050248 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050292 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050303 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050401 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050822 4706 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050851 4706 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050862 4706 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.050874 4706 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.061026 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.152204 4706 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.216143 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.619018 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.619617 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.630386 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.630524 4706 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="6a5321c3-1cb3-4462-8c86-46e1d35b9a1d" Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.633545 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 08 19:26:47 crc kubenswrapper[4706]: I1208 19:26:47.633601 4706 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="6a5321c3-1cb3-4462-8c86-46e1d35b9a1d" Dec 08 19:26:57 crc kubenswrapper[4706]: I1208 19:26:57.421937 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 08 19:26:57 crc kubenswrapper[4706]: I1208 19:26:57.487406 4706 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 08 19:26:57 crc kubenswrapper[4706]: I1208 19:26:57.928734 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 08 19:26:58 crc kubenswrapper[4706]: I1208 19:26:58.460495 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 08 19:27:02 crc kubenswrapper[4706]: I1208 19:27:02.679566 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.340614 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xwgxh"] Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.340971 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" podUID="979bd1cb-70e1-4f2d-b623-aaae40466dff" containerName="controller-manager" containerID="cri-o://1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614" gracePeriod=30 Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.438173 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q"] Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.438535 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" podUID="98220d47-1a06-4e17-90ec-f802f041197c" containerName="route-controller-manager" containerID="cri-o://dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d" gracePeriod=30 Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.723215 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.877823 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.886720 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdnwx\" (UniqueName: \"kubernetes.io/projected/979bd1cb-70e1-4f2d-b623-aaae40466dff-kube-api-access-mdnwx\") pod \"979bd1cb-70e1-4f2d-b623-aaae40466dff\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.886778 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/979bd1cb-70e1-4f2d-b623-aaae40466dff-serving-cert\") pod \"979bd1cb-70e1-4f2d-b623-aaae40466dff\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.886839 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-proxy-ca-bundles\") pod \"979bd1cb-70e1-4f2d-b623-aaae40466dff\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.886945 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-client-ca\") pod \"979bd1cb-70e1-4f2d-b623-aaae40466dff\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.886968 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-config\") pod \"979bd1cb-70e1-4f2d-b623-aaae40466dff\" (UID: \"979bd1cb-70e1-4f2d-b623-aaae40466dff\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.888034 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-client-ca" (OuterVolumeSpecName: "client-ca") pod "979bd1cb-70e1-4f2d-b623-aaae40466dff" (UID: "979bd1cb-70e1-4f2d-b623-aaae40466dff"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.888132 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-config" (OuterVolumeSpecName: "config") pod "979bd1cb-70e1-4f2d-b623-aaae40466dff" (UID: "979bd1cb-70e1-4f2d-b623-aaae40466dff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.888674 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "979bd1cb-70e1-4f2d-b623-aaae40466dff" (UID: "979bd1cb-70e1-4f2d-b623-aaae40466dff"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.893573 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979bd1cb-70e1-4f2d-b623-aaae40466dff-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "979bd1cb-70e1-4f2d-b623-aaae40466dff" (UID: "979bd1cb-70e1-4f2d-b623-aaae40466dff"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.895205 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/979bd1cb-70e1-4f2d-b623-aaae40466dff-kube-api-access-mdnwx" (OuterVolumeSpecName: "kube-api-access-mdnwx") pod "979bd1cb-70e1-4f2d-b623-aaae40466dff" (UID: "979bd1cb-70e1-4f2d-b623-aaae40466dff"). InnerVolumeSpecName "kube-api-access-mdnwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988082 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-config\") pod \"98220d47-1a06-4e17-90ec-f802f041197c\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988181 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-client-ca\") pod \"98220d47-1a06-4e17-90ec-f802f041197c\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988212 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr5bn\" (UniqueName: \"kubernetes.io/projected/98220d47-1a06-4e17-90ec-f802f041197c-kube-api-access-kr5bn\") pod \"98220d47-1a06-4e17-90ec-f802f041197c\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988243 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98220d47-1a06-4e17-90ec-f802f041197c-serving-cert\") pod \"98220d47-1a06-4e17-90ec-f802f041197c\" (UID: \"98220d47-1a06-4e17-90ec-f802f041197c\") " Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988542 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988555 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988567 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdnwx\" (UniqueName: \"kubernetes.io/projected/979bd1cb-70e1-4f2d-b623-aaae40466dff-kube-api-access-mdnwx\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988576 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/979bd1cb-70e1-4f2d-b623-aaae40466dff-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.988589 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979bd1cb-70e1-4f2d-b623-aaae40466dff-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.989540 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-config" (OuterVolumeSpecName: "config") pod "98220d47-1a06-4e17-90ec-f802f041197c" (UID: "98220d47-1a06-4e17-90ec-f802f041197c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.989795 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-client-ca" (OuterVolumeSpecName: "client-ca") pod "98220d47-1a06-4e17-90ec-f802f041197c" (UID: "98220d47-1a06-4e17-90ec-f802f041197c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.992497 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98220d47-1a06-4e17-90ec-f802f041197c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "98220d47-1a06-4e17-90ec-f802f041197c" (UID: "98220d47-1a06-4e17-90ec-f802f041197c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:27:03 crc kubenswrapper[4706]: I1208 19:27:03.992982 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98220d47-1a06-4e17-90ec-f802f041197c-kube-api-access-kr5bn" (OuterVolumeSpecName: "kube-api-access-kr5bn") pod "98220d47-1a06-4e17-90ec-f802f041197c" (UID: "98220d47-1a06-4e17-90ec-f802f041197c"). InnerVolumeSpecName "kube-api-access-kr5bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.089998 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.090037 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr5bn\" (UniqueName: \"kubernetes.io/projected/98220d47-1a06-4e17-90ec-f802f041197c-kube-api-access-kr5bn\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.090052 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98220d47-1a06-4e17-90ec-f802f041197c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.090062 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98220d47-1a06-4e17-90ec-f802f041197c-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.310333 4706 generic.go:334] "Generic (PLEG): container finished" podID="979bd1cb-70e1-4f2d-b623-aaae40466dff" containerID="1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614" exitCode=0 Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.310404 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.310436 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" event={"ID":"979bd1cb-70e1-4f2d-b623-aaae40466dff","Type":"ContainerDied","Data":"1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614"} Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.311068 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xwgxh" event={"ID":"979bd1cb-70e1-4f2d-b623-aaae40466dff","Type":"ContainerDied","Data":"26ecfa47e41147d5b5e8d62fab925acb3569a9e751b3db51bb748332c180de07"} Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.311107 4706 scope.go:117] "RemoveContainer" containerID="1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.315228 4706 generic.go:334] "Generic (PLEG): container finished" podID="98220d47-1a06-4e17-90ec-f802f041197c" containerID="dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d" exitCode=0 Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.315298 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" event={"ID":"98220d47-1a06-4e17-90ec-f802f041197c","Type":"ContainerDied","Data":"dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d"} Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.315329 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" event={"ID":"98220d47-1a06-4e17-90ec-f802f041197c","Type":"ContainerDied","Data":"bb5f1620b9e5f660c57c429c960fadaf5c1c3349cf92deb189808b821b9d2a89"} Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.315383 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.334237 4706 scope.go:117] "RemoveContainer" containerID="1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614" Dec 08 19:27:04 crc kubenswrapper[4706]: E1208 19:27:04.334796 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614\": container with ID starting with 1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614 not found: ID does not exist" containerID="1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.334838 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614"} err="failed to get container status \"1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614\": rpc error: code = NotFound desc = could not find container \"1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614\": container with ID starting with 1a885828a2ec9f672da55b3fe413c210aefb962b0b76562f8f58b23cd0523614 not found: ID does not exist" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.334866 4706 scope.go:117] "RemoveContainer" containerID="dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.360502 4706 scope.go:117] "RemoveContainer" containerID="dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.360730 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q"] Dec 08 19:27:04 crc kubenswrapper[4706]: E1208 19:27:04.360991 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d\": container with ID starting with dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d not found: ID does not exist" containerID="dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.361030 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d"} err="failed to get container status \"dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d\": rpc error: code = NotFound desc = could not find container \"dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d\": container with ID starting with dea86ac6d2347f376683a896b4099288ef4e44b0830b306b1433b5519e03b17d not found: ID does not exist" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.365193 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pk89q"] Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.369475 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xwgxh"] Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.373552 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xwgxh"] Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.687689 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.888202 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.937638 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-599b86c796-lpb2h"] Dec 08 19:27:04 crc kubenswrapper[4706]: E1208 19:27:04.938566 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98220d47-1a06-4e17-90ec-f802f041197c" containerName="route-controller-manager" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.938588 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="98220d47-1a06-4e17-90ec-f802f041197c" containerName="route-controller-manager" Dec 08 19:27:04 crc kubenswrapper[4706]: E1208 19:27:04.938623 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.938633 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 08 19:27:04 crc kubenswrapper[4706]: E1208 19:27:04.938650 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="979bd1cb-70e1-4f2d-b623-aaae40466dff" containerName="controller-manager" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.938662 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="979bd1cb-70e1-4f2d-b623-aaae40466dff" containerName="controller-manager" Dec 08 19:27:04 crc kubenswrapper[4706]: E1208 19:27:04.938673 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" containerName="installer" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.938687 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" containerName="installer" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.938881 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.938902 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="98220d47-1a06-4e17-90ec-f802f041197c" containerName="route-controller-manager" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.938914 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="979bd1cb-70e1-4f2d-b623-aaae40466dff" containerName="controller-manager" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.938926 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d87b9a0-63a8-4619-88cb-127f2608edd6" containerName="installer" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.939642 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.943055 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.944689 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.944826 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.944908 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.944826 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.945766 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.949466 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj"] Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.950539 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.954839 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.955143 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.955715 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.956084 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.956301 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.957034 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.957623 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj"] Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.960199 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 08 19:27:04 crc kubenswrapper[4706]: I1208 19:27:04.963391 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-599b86c796-lpb2h"] Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.106553 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-client-ca\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.106613 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-serving-cert\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.106647 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-config\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.106685 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-client-ca\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.106715 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xg69\" (UniqueName: \"kubernetes.io/projected/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-kube-api-access-8xg69\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.106943 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-config\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.107128 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzh6k\" (UniqueName: \"kubernetes.io/projected/52fe9600-233f-4c54-a6db-38d2a5e62d37-kube-api-access-hzh6k\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.107179 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-proxy-ca-bundles\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.107297 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52fe9600-233f-4c54-a6db-38d2a5e62d37-serving-cert\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209059 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzh6k\" (UniqueName: \"kubernetes.io/projected/52fe9600-233f-4c54-a6db-38d2a5e62d37-kube-api-access-hzh6k\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209139 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-proxy-ca-bundles\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209190 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52fe9600-233f-4c54-a6db-38d2a5e62d37-serving-cert\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209232 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-client-ca\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209279 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-serving-cert\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209318 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-config\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209369 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-client-ca\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209395 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xg69\" (UniqueName: \"kubernetes.io/projected/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-kube-api-access-8xg69\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.209423 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-config\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.210398 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-client-ca\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.210973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-config\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.211100 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-proxy-ca-bundles\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.211535 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-client-ca\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.211682 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-config\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.216103 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-serving-cert\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.220295 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52fe9600-233f-4c54-a6db-38d2a5e62d37-serving-cert\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.227315 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xg69\" (UniqueName: \"kubernetes.io/projected/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-kube-api-access-8xg69\") pod \"route-controller-manager-86dcf56f75-shxkj\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.232146 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzh6k\" (UniqueName: \"kubernetes.io/projected/52fe9600-233f-4c54-a6db-38d2a5e62d37-kube-api-access-hzh6k\") pod \"controller-manager-599b86c796-lpb2h\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.265232 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.282929 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.461109 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-599b86c796-lpb2h"] Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.513652 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj"] Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.624744 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="979bd1cb-70e1-4f2d-b623-aaae40466dff" path="/var/lib/kubelet/pods/979bd1cb-70e1-4f2d-b623-aaae40466dff/volumes" Dec 08 19:27:05 crc kubenswrapper[4706]: I1208 19:27:05.625702 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98220d47-1a06-4e17-90ec-f802f041197c" path="/var/lib/kubelet/pods/98220d47-1a06-4e17-90ec-f802f041197c/volumes" Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.335789 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" event={"ID":"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4","Type":"ContainerStarted","Data":"8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0"} Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.336166 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.336184 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" event={"ID":"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4","Type":"ContainerStarted","Data":"b9d18353ee2bee6006f6b737b7c4ae9749267073b6d38afcd35227a68bbed304"} Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.337470 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" event={"ID":"52fe9600-233f-4c54-a6db-38d2a5e62d37","Type":"ContainerStarted","Data":"1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442"} Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.337520 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" event={"ID":"52fe9600-233f-4c54-a6db-38d2a5e62d37","Type":"ContainerStarted","Data":"d04e9018206c023c5fc8df5f2bc04d53b3d28b1b45a1f514b3b206138559299a"} Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.338128 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.345988 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.358463 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.365423 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" podStartSLOduration=3.3654037949999998 podStartE2EDuration="3.365403795s" podCreationTimestamp="2025-12-08 19:27:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:27:06.364396633 +0000 UTC m=+309.006597636" watchObservedRunningTime="2025-12-08 19:27:06.365403795 +0000 UTC m=+309.007604798" Dec 08 19:27:06 crc kubenswrapper[4706]: I1208 19:27:06.442567 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" podStartSLOduration=3.442541097 podStartE2EDuration="3.442541097s" podCreationTimestamp="2025-12-08 19:27:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:27:06.437067603 +0000 UTC m=+309.079268606" watchObservedRunningTime="2025-12-08 19:27:06.442541097 +0000 UTC m=+309.084742100" Dec 08 19:27:07 crc kubenswrapper[4706]: I1208 19:27:07.328811 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 08 19:27:07 crc kubenswrapper[4706]: I1208 19:27:07.623475 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 08 19:27:08 crc kubenswrapper[4706]: I1208 19:27:08.956722 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.135240 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cqfh4"] Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.137742 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.139946 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.146780 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cqfh4"] Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.285782 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584d998e-f5d2-40eb-98d9-a3d5a6918144-catalog-content\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.285860 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74dbb\" (UniqueName: \"kubernetes.io/projected/584d998e-f5d2-40eb-98d9-a3d5a6918144-kube-api-access-74dbb\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.285905 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584d998e-f5d2-40eb-98d9-a3d5a6918144-utilities\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.387385 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584d998e-f5d2-40eb-98d9-a3d5a6918144-catalog-content\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.387489 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74dbb\" (UniqueName: \"kubernetes.io/projected/584d998e-f5d2-40eb-98d9-a3d5a6918144-kube-api-access-74dbb\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.387544 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584d998e-f5d2-40eb-98d9-a3d5a6918144-utilities\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.388102 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584d998e-f5d2-40eb-98d9-a3d5a6918144-catalog-content\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.388251 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584d998e-f5d2-40eb-98d9-a3d5a6918144-utilities\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.407110 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74dbb\" (UniqueName: \"kubernetes.io/projected/584d998e-f5d2-40eb-98d9-a3d5a6918144-kube-api-access-74dbb\") pod \"community-operators-cqfh4\" (UID: \"584d998e-f5d2-40eb-98d9-a3d5a6918144\") " pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.456605 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.824695 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 08 19:27:12 crc kubenswrapper[4706]: I1208 19:27:12.857885 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cqfh4"] Dec 08 19:27:13 crc kubenswrapper[4706]: I1208 19:27:13.379034 4706 generic.go:334] "Generic (PLEG): container finished" podID="584d998e-f5d2-40eb-98d9-a3d5a6918144" containerID="30c6e1c3d41a6fb260b73edbaae77ec8e1beb5a72a308f06de79d825123f2cc1" exitCode=0 Dec 08 19:27:13 crc kubenswrapper[4706]: I1208 19:27:13.379083 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqfh4" event={"ID":"584d998e-f5d2-40eb-98d9-a3d5a6918144","Type":"ContainerDied","Data":"30c6e1c3d41a6fb260b73edbaae77ec8e1beb5a72a308f06de79d825123f2cc1"} Dec 08 19:27:13 crc kubenswrapper[4706]: I1208 19:27:13.379111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqfh4" event={"ID":"584d998e-f5d2-40eb-98d9-a3d5a6918144","Type":"ContainerStarted","Data":"af5c9b8057c8be77a10933b48cf70c4c86b1692ce865844625883d2b2ef921e0"} Dec 08 19:27:14 crc kubenswrapper[4706]: I1208 19:27:14.387379 4706 generic.go:334] "Generic (PLEG): container finished" podID="584d998e-f5d2-40eb-98d9-a3d5a6918144" containerID="d249dcaba09e6142387c56e0f387fc920fe550ed750575a51a296ffb52782b64" exitCode=0 Dec 08 19:27:14 crc kubenswrapper[4706]: I1208 19:27:14.387450 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqfh4" event={"ID":"584d998e-f5d2-40eb-98d9-a3d5a6918144","Type":"ContainerDied","Data":"d249dcaba09e6142387c56e0f387fc920fe550ed750575a51a296ffb52782b64"} Dec 08 19:27:15 crc kubenswrapper[4706]: I1208 19:27:15.164384 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 08 19:27:15 crc kubenswrapper[4706]: I1208 19:27:15.393724 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 08 19:27:15 crc kubenswrapper[4706]: I1208 19:27:15.395561 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqfh4" event={"ID":"584d998e-f5d2-40eb-98d9-a3d5a6918144","Type":"ContainerStarted","Data":"07e2418f78589c87df7562dc020b60f3d757cdc6d836f6916b6e721412e0b51d"} Dec 08 19:27:15 crc kubenswrapper[4706]: I1208 19:27:15.416740 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cqfh4" podStartSLOduration=2.027128905 podStartE2EDuration="3.416716313s" podCreationTimestamp="2025-12-08 19:27:12 +0000 UTC" firstStartedPulling="2025-12-08 19:27:13.380589915 +0000 UTC m=+316.022790938" lastFinishedPulling="2025-12-08 19:27:14.770177343 +0000 UTC m=+317.412378346" observedRunningTime="2025-12-08 19:27:15.414290626 +0000 UTC m=+318.056491639" watchObservedRunningTime="2025-12-08 19:27:15.416716313 +0000 UTC m=+318.058917306" Dec 08 19:27:16 crc kubenswrapper[4706]: I1208 19:27:16.562769 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 08 19:27:20 crc kubenswrapper[4706]: I1208 19:27:20.816565 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 08 19:27:22 crc kubenswrapper[4706]: I1208 19:27:22.247192 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 08 19:27:22 crc kubenswrapper[4706]: I1208 19:27:22.457550 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:22 crc kubenswrapper[4706]: I1208 19:27:22.457629 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:22 crc kubenswrapper[4706]: I1208 19:27:22.516563 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.326892 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mj8dj"] Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.328460 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.330885 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.334376 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mj8dj"] Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.446971 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a2b5a45-e2a5-4793-b284-38116ab1b244-utilities\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.447033 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mncdp\" (UniqueName: \"kubernetes.io/projected/4a2b5a45-e2a5-4793-b284-38116ab1b244-kube-api-access-mncdp\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.447766 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a2b5a45-e2a5-4793-b284-38116ab1b244-catalog-content\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.489412 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cqfh4" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.548775 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a2b5a45-e2a5-4793-b284-38116ab1b244-catalog-content\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.549833 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a2b5a45-e2a5-4793-b284-38116ab1b244-catalog-content\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.550018 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a2b5a45-e2a5-4793-b284-38116ab1b244-utilities\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.550336 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a2b5a45-e2a5-4793-b284-38116ab1b244-utilities\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.550386 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mncdp\" (UniqueName: \"kubernetes.io/projected/4a2b5a45-e2a5-4793-b284-38116ab1b244-kube-api-access-mncdp\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.583025 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mncdp\" (UniqueName: \"kubernetes.io/projected/4a2b5a45-e2a5-4793-b284-38116ab1b244-kube-api-access-mncdp\") pod \"certified-operators-mj8dj\" (UID: \"4a2b5a45-e2a5-4793-b284-38116ab1b244\") " pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:23 crc kubenswrapper[4706]: I1208 19:27:23.656206 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.041953 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mj8dj"] Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.454751 4706 generic.go:334] "Generic (PLEG): container finished" podID="4a2b5a45-e2a5-4793-b284-38116ab1b244" containerID="b42454e70541ac4a594ce0b3d50316ee7bdbfa95a4fdfd0ae636dd5919f26af5" exitCode=0 Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.454865 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mj8dj" event={"ID":"4a2b5a45-e2a5-4793-b284-38116ab1b244","Type":"ContainerDied","Data":"b42454e70541ac4a594ce0b3d50316ee7bdbfa95a4fdfd0ae636dd5919f26af5"} Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.454951 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mj8dj" event={"ID":"4a2b5a45-e2a5-4793-b284-38116ab1b244","Type":"ContainerStarted","Data":"5fa26505eb2c52b31f6ba6b59ddc5fe8d629f7a0f0deac03bf9adaf88b1bf0d8"} Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.707945 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s5pmf"] Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.709329 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.711970 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.721083 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s5pmf"] Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.867057 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-catalog-content\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.867618 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w4bx\" (UniqueName: \"kubernetes.io/projected/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-kube-api-access-9w4bx\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.867768 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-utilities\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.969412 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w4bx\" (UniqueName: \"kubernetes.io/projected/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-kube-api-access-9w4bx\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.969523 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-utilities\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.969559 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-catalog-content\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.970292 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-catalog-content\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.970409 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-utilities\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:24 crc kubenswrapper[4706]: I1208 19:27:24.999464 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w4bx\" (UniqueName: \"kubernetes.io/projected/c4f715e7-a6b1-4f7d-b1fe-cdc455423864-kube-api-access-9w4bx\") pod \"redhat-marketplace-s5pmf\" (UID: \"c4f715e7-a6b1-4f7d-b1fe-cdc455423864\") " pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.038705 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.241022 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.452739 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s5pmf"] Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.465783 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mj8dj" event={"ID":"4a2b5a45-e2a5-4793-b284-38116ab1b244","Type":"ContainerStarted","Data":"d9be47f7a2483a1eab054362e58429dff9fb1d38f6e29c018c28b06831de5b7f"} Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.708697 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l2crj"] Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.714892 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.718435 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.741652 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l2crj"] Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.886137 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct7fn\" (UniqueName: \"kubernetes.io/projected/da615802-59e8-4f61-9543-cd20ced42c39-kube-api-access-ct7fn\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.886214 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da615802-59e8-4f61-9543-cd20ced42c39-utilities\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.886277 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da615802-59e8-4f61-9543-cd20ced42c39-catalog-content\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.987898 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da615802-59e8-4f61-9543-cd20ced42c39-catalog-content\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.987982 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct7fn\" (UniqueName: \"kubernetes.io/projected/da615802-59e8-4f61-9543-cd20ced42c39-kube-api-access-ct7fn\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.988020 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da615802-59e8-4f61-9543-cd20ced42c39-utilities\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.988521 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da615802-59e8-4f61-9543-cd20ced42c39-catalog-content\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:25 crc kubenswrapper[4706]: I1208 19:27:25.988532 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da615802-59e8-4f61-9543-cd20ced42c39-utilities\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.008780 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct7fn\" (UniqueName: \"kubernetes.io/projected/da615802-59e8-4f61-9543-cd20ced42c39-kube-api-access-ct7fn\") pod \"redhat-operators-l2crj\" (UID: \"da615802-59e8-4f61-9543-cd20ced42c39\") " pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.039366 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.449193 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l2crj"] Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.479108 4706 generic.go:334] "Generic (PLEG): container finished" podID="c4f715e7-a6b1-4f7d-b1fe-cdc455423864" containerID="87577686d91081f02ff0d1aa0a3c9e14e5d9e750c63d4d41d18737d8edea5997" exitCode=0 Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.479177 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5pmf" event={"ID":"c4f715e7-a6b1-4f7d-b1fe-cdc455423864","Type":"ContainerDied","Data":"87577686d91081f02ff0d1aa0a3c9e14e5d9e750c63d4d41d18737d8edea5997"} Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.479206 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5pmf" event={"ID":"c4f715e7-a6b1-4f7d-b1fe-cdc455423864","Type":"ContainerStarted","Data":"e147ee9341bf5b422c81e92df1a6f21db9e183a74b1c4987229470dcbdf700ed"} Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.489001 4706 generic.go:334] "Generic (PLEG): container finished" podID="4a2b5a45-e2a5-4793-b284-38116ab1b244" containerID="d9be47f7a2483a1eab054362e58429dff9fb1d38f6e29c018c28b06831de5b7f" exitCode=0 Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.489111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mj8dj" event={"ID":"4a2b5a45-e2a5-4793-b284-38116ab1b244","Type":"ContainerDied","Data":"d9be47f7a2483a1eab054362e58429dff9fb1d38f6e29c018c28b06831de5b7f"} Dec 08 19:27:26 crc kubenswrapper[4706]: I1208 19:27:26.491143 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2crj" event={"ID":"da615802-59e8-4f61-9543-cd20ced42c39","Type":"ContainerStarted","Data":"88d1f4db0af662913fa37f594284a466d5ee72d3bb7e5df189ab81df93c05387"} Dec 08 19:27:27 crc kubenswrapper[4706]: I1208 19:27:27.498958 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mj8dj" event={"ID":"4a2b5a45-e2a5-4793-b284-38116ab1b244","Type":"ContainerStarted","Data":"dfe17e3e24b0dd3cf946c303ad54c2274d462ae012119647e52c0b4bdfcad2f3"} Dec 08 19:27:27 crc kubenswrapper[4706]: I1208 19:27:27.501663 4706 generic.go:334] "Generic (PLEG): container finished" podID="da615802-59e8-4f61-9543-cd20ced42c39" containerID="eb1bf82ad30f6d577f1157aa02f1886fc603bc120d6d78724a42ec953c97983d" exitCode=0 Dec 08 19:27:27 crc kubenswrapper[4706]: I1208 19:27:27.501724 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2crj" event={"ID":"da615802-59e8-4f61-9543-cd20ced42c39","Type":"ContainerDied","Data":"eb1bf82ad30f6d577f1157aa02f1886fc603bc120d6d78724a42ec953c97983d"} Dec 08 19:27:27 crc kubenswrapper[4706]: I1208 19:27:27.504333 4706 generic.go:334] "Generic (PLEG): container finished" podID="c4f715e7-a6b1-4f7d-b1fe-cdc455423864" containerID="e3d8892f18c3b7a20dc82f4cdb33c58931369673df281eb364a57711cbd16fca" exitCode=0 Dec 08 19:27:27 crc kubenswrapper[4706]: I1208 19:27:27.504384 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5pmf" event={"ID":"c4f715e7-a6b1-4f7d-b1fe-cdc455423864","Type":"ContainerDied","Data":"e3d8892f18c3b7a20dc82f4cdb33c58931369673df281eb364a57711cbd16fca"} Dec 08 19:27:27 crc kubenswrapper[4706]: I1208 19:27:27.521663 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mj8dj" podStartSLOduration=2.088980359 podStartE2EDuration="4.521640297s" podCreationTimestamp="2025-12-08 19:27:23 +0000 UTC" firstStartedPulling="2025-12-08 19:27:24.457230818 +0000 UTC m=+327.099431891" lastFinishedPulling="2025-12-08 19:27:26.889890826 +0000 UTC m=+329.532091829" observedRunningTime="2025-12-08 19:27:27.517449691 +0000 UTC m=+330.159650704" watchObservedRunningTime="2025-12-08 19:27:27.521640297 +0000 UTC m=+330.163841300" Dec 08 19:27:28 crc kubenswrapper[4706]: I1208 19:27:28.511355 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2crj" event={"ID":"da615802-59e8-4f61-9543-cd20ced42c39","Type":"ContainerStarted","Data":"e8f715e717fbd10b6246f42b7aeb14d7c3e9463d2f6a6b36e28c5becce4d691d"} Dec 08 19:27:28 crc kubenswrapper[4706]: I1208 19:27:28.513528 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5pmf" event={"ID":"c4f715e7-a6b1-4f7d-b1fe-cdc455423864","Type":"ContainerStarted","Data":"4755a8535efc1981bc645d359a0988528a112035269c3a1f18473c8b71dea4b2"} Dec 08 19:27:28 crc kubenswrapper[4706]: I1208 19:27:28.553728 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s5pmf" podStartSLOduration=3.143118 podStartE2EDuration="4.553699386s" podCreationTimestamp="2025-12-08 19:27:24 +0000 UTC" firstStartedPulling="2025-12-08 19:27:26.481454254 +0000 UTC m=+329.123655257" lastFinishedPulling="2025-12-08 19:27:27.89203564 +0000 UTC m=+330.534236643" observedRunningTime="2025-12-08 19:27:28.549415748 +0000 UTC m=+331.191616751" watchObservedRunningTime="2025-12-08 19:27:28.553699386 +0000 UTC m=+331.195900389" Dec 08 19:27:29 crc kubenswrapper[4706]: I1208 19:27:29.520853 4706 generic.go:334] "Generic (PLEG): container finished" podID="da615802-59e8-4f61-9543-cd20ced42c39" containerID="e8f715e717fbd10b6246f42b7aeb14d7c3e9463d2f6a6b36e28c5becce4d691d" exitCode=0 Dec 08 19:27:29 crc kubenswrapper[4706]: I1208 19:27:29.520925 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2crj" event={"ID":"da615802-59e8-4f61-9543-cd20ced42c39","Type":"ContainerDied","Data":"e8f715e717fbd10b6246f42b7aeb14d7c3e9463d2f6a6b36e28c5becce4d691d"} Dec 08 19:27:30 crc kubenswrapper[4706]: I1208 19:27:30.529772 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2crj" event={"ID":"da615802-59e8-4f61-9543-cd20ced42c39","Type":"ContainerStarted","Data":"732c628a8a3ce343be48c6417b41be29956228eb17169942daf2538030eabe7e"} Dec 08 19:27:30 crc kubenswrapper[4706]: I1208 19:27:30.548480 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l2crj" podStartSLOduration=3.115433631 podStartE2EDuration="5.548454397s" podCreationTimestamp="2025-12-08 19:27:25 +0000 UTC" firstStartedPulling="2025-12-08 19:27:27.50371159 +0000 UTC m=+330.145912593" lastFinishedPulling="2025-12-08 19:27:29.936732356 +0000 UTC m=+332.578933359" observedRunningTime="2025-12-08 19:27:30.548313443 +0000 UTC m=+333.190514456" watchObservedRunningTime="2025-12-08 19:27:30.548454397 +0000 UTC m=+333.190655400" Dec 08 19:27:33 crc kubenswrapper[4706]: I1208 19:27:33.657192 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:33 crc kubenswrapper[4706]: I1208 19:27:33.658786 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:33 crc kubenswrapper[4706]: I1208 19:27:33.700390 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:34 crc kubenswrapper[4706]: I1208 19:27:34.592803 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mj8dj" Dec 08 19:27:35 crc kubenswrapper[4706]: I1208 19:27:35.038983 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:35 crc kubenswrapper[4706]: I1208 19:27:35.039074 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:35 crc kubenswrapper[4706]: I1208 19:27:35.075906 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:35 crc kubenswrapper[4706]: I1208 19:27:35.640012 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s5pmf" Dec 08 19:27:36 crc kubenswrapper[4706]: I1208 19:27:36.039967 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:36 crc kubenswrapper[4706]: I1208 19:27:36.040838 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:36 crc kubenswrapper[4706]: I1208 19:27:36.088390 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:36 crc kubenswrapper[4706]: I1208 19:27:36.610371 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l2crj" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.562123 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k9b7x"] Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.563658 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.579846 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k9b7x"] Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.710417 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/077616e4-3a8a-4c48-bac8-c63d04462409-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.710487 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/077616e4-3a8a-4c48-bac8-c63d04462409-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.710723 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnmqd\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-kube-api-access-tnmqd\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.710844 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-registry-tls\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.710966 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/077616e4-3a8a-4c48-bac8-c63d04462409-trusted-ca\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.711052 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.711098 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/077616e4-3a8a-4c48-bac8-c63d04462409-registry-certificates\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.711162 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-bound-sa-token\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.742510 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.813218 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/077616e4-3a8a-4c48-bac8-c63d04462409-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.813306 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/077616e4-3a8a-4c48-bac8-c63d04462409-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.813341 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnmqd\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-kube-api-access-tnmqd\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.813377 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-registry-tls\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.813410 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/077616e4-3a8a-4c48-bac8-c63d04462409-trusted-ca\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.813442 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/077616e4-3a8a-4c48-bac8-c63d04462409-registry-certificates\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.813486 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-bound-sa-token\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.814208 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/077616e4-3a8a-4c48-bac8-c63d04462409-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.816237 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/077616e4-3a8a-4c48-bac8-c63d04462409-trusted-ca\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.819332 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/077616e4-3a8a-4c48-bac8-c63d04462409-registry-certificates\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.821119 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/077616e4-3a8a-4c48-bac8-c63d04462409-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.821124 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-registry-tls\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.834372 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-bound-sa-token\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.835233 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnmqd\" (UniqueName: \"kubernetes.io/projected/077616e4-3a8a-4c48-bac8-c63d04462409-kube-api-access-tnmqd\") pod \"image-registry-66df7c8f76-k9b7x\" (UID: \"077616e4-3a8a-4c48-bac8-c63d04462409\") " pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:49 crc kubenswrapper[4706]: I1208 19:27:49.881212 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:50 crc kubenswrapper[4706]: I1208 19:27:50.282892 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k9b7x"] Dec 08 19:27:50 crc kubenswrapper[4706]: I1208 19:27:50.665012 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" event={"ID":"077616e4-3a8a-4c48-bac8-c63d04462409","Type":"ContainerStarted","Data":"eb896b9f34c7841ad45782154cd8dd576ce00ba2518675b32e2d3019af3a2c45"} Dec 08 19:27:51 crc kubenswrapper[4706]: I1208 19:27:51.679776 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" event={"ID":"077616e4-3a8a-4c48-bac8-c63d04462409","Type":"ContainerStarted","Data":"c0c54cb13a47eda0ef45dfda371762b99f9ccc46b8bd997a4c001235b326d578"} Dec 08 19:27:51 crc kubenswrapper[4706]: I1208 19:27:51.680541 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:27:51 crc kubenswrapper[4706]: I1208 19:27:51.707697 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" podStartSLOduration=2.7076804230000002 podStartE2EDuration="2.707680423s" podCreationTimestamp="2025-12-08 19:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:27:51.704128217 +0000 UTC m=+354.346329220" watchObservedRunningTime="2025-12-08 19:27:51.707680423 +0000 UTC m=+354.349881426" Dec 08 19:28:05 crc kubenswrapper[4706]: I1208 19:28:05.836370 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:28:05 crc kubenswrapper[4706]: I1208 19:28:05.837056 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:28:09 crc kubenswrapper[4706]: I1208 19:28:09.892702 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-k9b7x" Dec 08 19:28:09 crc kubenswrapper[4706]: I1208 19:28:09.989208 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-28lq2"] Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.381649 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj"] Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.382697 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" podUID="7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" containerName="route-controller-manager" containerID="cri-o://8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0" gracePeriod=30 Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.389699 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-599b86c796-lpb2h"] Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.390044 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" podUID="52fe9600-233f-4c54-a6db-38d2a5e62d37" containerName="controller-manager" containerID="cri-o://1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442" gracePeriod=30 Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.795969 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.796930 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.866488 4706 generic.go:334] "Generic (PLEG): container finished" podID="7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" containerID="8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0" exitCode=0 Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.866582 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.866609 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" event={"ID":"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4","Type":"ContainerDied","Data":"8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0"} Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.866689 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj" event={"ID":"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4","Type":"ContainerDied","Data":"b9d18353ee2bee6006f6b737b7c4ae9749267073b6d38afcd35227a68bbed304"} Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.866712 4706 scope.go:117] "RemoveContainer" containerID="8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.870817 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.871052 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" event={"ID":"52fe9600-233f-4c54-a6db-38d2a5e62d37","Type":"ContainerDied","Data":"1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442"} Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.870591 4706 generic.go:334] "Generic (PLEG): container finished" podID="52fe9600-233f-4c54-a6db-38d2a5e62d37" containerID="1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442" exitCode=0 Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.871252 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-599b86c796-lpb2h" event={"ID":"52fe9600-233f-4c54-a6db-38d2a5e62d37","Type":"ContainerDied","Data":"d04e9018206c023c5fc8df5f2bc04d53b3d28b1b45a1f514b3b206138559299a"} Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.885290 4706 scope.go:117] "RemoveContainer" containerID="8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0" Dec 08 19:28:23 crc kubenswrapper[4706]: E1208 19:28:23.885716 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0\": container with ID starting with 8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0 not found: ID does not exist" containerID="8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.885754 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0"} err="failed to get container status \"8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0\": rpc error: code = NotFound desc = could not find container \"8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0\": container with ID starting with 8fad6b5e0e91e95cbdbf637e8ed5d40e7ffd4e43b1ec7f50414f648e22b9def0 not found: ID does not exist" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.885778 4706 scope.go:117] "RemoveContainer" containerID="1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.900619 4706 scope.go:117] "RemoveContainer" containerID="1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442" Dec 08 19:28:23 crc kubenswrapper[4706]: E1208 19:28:23.901019 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442\": container with ID starting with 1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442 not found: ID does not exist" containerID="1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.901084 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442"} err="failed to get container status \"1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442\": rpc error: code = NotFound desc = could not find container \"1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442\": container with ID starting with 1e32e5392eb1de2f5cd1eb108e14f9e2458ddf9ff9e14ad9f8993127de5ba442 not found: ID does not exist" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.909730 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xg69\" (UniqueName: \"kubernetes.io/projected/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-kube-api-access-8xg69\") pod \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.909783 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-client-ca\") pod \"52fe9600-233f-4c54-a6db-38d2a5e62d37\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.909855 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzh6k\" (UniqueName: \"kubernetes.io/projected/52fe9600-233f-4c54-a6db-38d2a5e62d37-kube-api-access-hzh6k\") pod \"52fe9600-233f-4c54-a6db-38d2a5e62d37\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.909927 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52fe9600-233f-4c54-a6db-38d2a5e62d37-serving-cert\") pod \"52fe9600-233f-4c54-a6db-38d2a5e62d37\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.909977 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-client-ca\") pod \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.910003 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-config\") pod \"52fe9600-233f-4c54-a6db-38d2a5e62d37\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.910023 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-config\") pod \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.910054 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-serving-cert\") pod \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\" (UID: \"7e72f6a7-027d-4327-b9eb-39e28a4ab2e4\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.910071 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-proxy-ca-bundles\") pod \"52fe9600-233f-4c54-a6db-38d2a5e62d37\" (UID: \"52fe9600-233f-4c54-a6db-38d2a5e62d37\") " Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.911447 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-client-ca" (OuterVolumeSpecName: "client-ca") pod "7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" (UID: "7e72f6a7-027d-4327-b9eb-39e28a4ab2e4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.911641 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-config" (OuterVolumeSpecName: "config") pod "7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" (UID: "7e72f6a7-027d-4327-b9eb-39e28a4ab2e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.911685 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-client-ca" (OuterVolumeSpecName: "client-ca") pod "52fe9600-233f-4c54-a6db-38d2a5e62d37" (UID: "52fe9600-233f-4c54-a6db-38d2a5e62d37"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.911706 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "52fe9600-233f-4c54-a6db-38d2a5e62d37" (UID: "52fe9600-233f-4c54-a6db-38d2a5e62d37"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.911819 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-config" (OuterVolumeSpecName: "config") pod "52fe9600-233f-4c54-a6db-38d2a5e62d37" (UID: "52fe9600-233f-4c54-a6db-38d2a5e62d37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.917601 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" (UID: "7e72f6a7-027d-4327-b9eb-39e28a4ab2e4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.917982 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52fe9600-233f-4c54-a6db-38d2a5e62d37-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "52fe9600-233f-4c54-a6db-38d2a5e62d37" (UID: "52fe9600-233f-4c54-a6db-38d2a5e62d37"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.917990 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-kube-api-access-8xg69" (OuterVolumeSpecName: "kube-api-access-8xg69") pod "7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" (UID: "7e72f6a7-027d-4327-b9eb-39e28a4ab2e4"). InnerVolumeSpecName "kube-api-access-8xg69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:28:23 crc kubenswrapper[4706]: I1208 19:28:23.918057 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52fe9600-233f-4c54-a6db-38d2a5e62d37-kube-api-access-hzh6k" (OuterVolumeSpecName: "kube-api-access-hzh6k") pod "52fe9600-233f-4c54-a6db-38d2a5e62d37" (UID: "52fe9600-233f-4c54-a6db-38d2a5e62d37"). InnerVolumeSpecName "kube-api-access-hzh6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011635 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52fe9600-233f-4c54-a6db-38d2a5e62d37-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011686 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011703 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011716 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011728 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011739 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011756 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xg69\" (UniqueName: \"kubernetes.io/projected/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4-kube-api-access-8xg69\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011770 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52fe9600-233f-4c54-a6db-38d2a5e62d37-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.011783 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzh6k\" (UniqueName: \"kubernetes.io/projected/52fe9600-233f-4c54-a6db-38d2a5e62d37-kube-api-access-hzh6k\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.213962 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj"] Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.220459 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-shxkj"] Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.235222 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-599b86c796-lpb2h"] Dec 08 19:28:24 crc kubenswrapper[4706]: I1208 19:28:24.236173 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-599b86c796-lpb2h"] Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.004190 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp"] Dec 08 19:28:25 crc kubenswrapper[4706]: E1208 19:28:25.004732 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" containerName="route-controller-manager" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.004764 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" containerName="route-controller-manager" Dec 08 19:28:25 crc kubenswrapper[4706]: E1208 19:28:25.004804 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52fe9600-233f-4c54-a6db-38d2a5e62d37" containerName="controller-manager" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.004820 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="52fe9600-233f-4c54-a6db-38d2a5e62d37" containerName="controller-manager" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.005060 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" containerName="route-controller-manager" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.005097 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="52fe9600-233f-4c54-a6db-38d2a5e62d37" containerName="controller-manager" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.005922 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.009573 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.009603 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.009655 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.009750 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.009784 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.009997 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.012329 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57968d6747-922m8"] Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.013616 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.015217 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.015297 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.015650 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.017471 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.017700 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp"] Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.017755 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.018096 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.021330 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57968d6747-922m8"] Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.027489 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.126976 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-config\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.127028 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-client-ca\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.127107 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89bf1c31-1caf-4a23-87e5-bbdba71df427-config\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.127308 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89bf1c31-1caf-4a23-87e5-bbdba71df427-client-ca\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.127378 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k757f\" (UniqueName: \"kubernetes.io/projected/89bf1c31-1caf-4a23-87e5-bbdba71df427-kube-api-access-k757f\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.127416 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3caf0864-a852-4a45-b426-d32b49dd4aa7-serving-cert\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.127494 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdk4c\" (UniqueName: \"kubernetes.io/projected/3caf0864-a852-4a45-b426-d32b49dd4aa7-kube-api-access-gdk4c\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.127537 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89bf1c31-1caf-4a23-87e5-bbdba71df427-serving-cert\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.127577 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-proxy-ca-bundles\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229305 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89bf1c31-1caf-4a23-87e5-bbdba71df427-client-ca\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229368 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k757f\" (UniqueName: \"kubernetes.io/projected/89bf1c31-1caf-4a23-87e5-bbdba71df427-kube-api-access-k757f\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229390 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3caf0864-a852-4a45-b426-d32b49dd4aa7-serving-cert\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229423 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdk4c\" (UniqueName: \"kubernetes.io/projected/3caf0864-a852-4a45-b426-d32b49dd4aa7-kube-api-access-gdk4c\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229447 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89bf1c31-1caf-4a23-87e5-bbdba71df427-serving-cert\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-proxy-ca-bundles\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229496 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-config\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229517 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-client-ca\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.229550 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89bf1c31-1caf-4a23-87e5-bbdba71df427-config\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.230399 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89bf1c31-1caf-4a23-87e5-bbdba71df427-client-ca\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.230688 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89bf1c31-1caf-4a23-87e5-bbdba71df427-config\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.231189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-client-ca\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.231400 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-config\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.232116 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3caf0864-a852-4a45-b426-d32b49dd4aa7-proxy-ca-bundles\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.238791 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3caf0864-a852-4a45-b426-d32b49dd4aa7-serving-cert\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.239752 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89bf1c31-1caf-4a23-87e5-bbdba71df427-serving-cert\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.247129 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdk4c\" (UniqueName: \"kubernetes.io/projected/3caf0864-a852-4a45-b426-d32b49dd4aa7-kube-api-access-gdk4c\") pod \"controller-manager-57968d6747-922m8\" (UID: \"3caf0864-a852-4a45-b426-d32b49dd4aa7\") " pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.248602 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k757f\" (UniqueName: \"kubernetes.io/projected/89bf1c31-1caf-4a23-87e5-bbdba71df427-kube-api-access-k757f\") pod \"route-controller-manager-76c99cf7b7-qcllp\" (UID: \"89bf1c31-1caf-4a23-87e5-bbdba71df427\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.342360 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.349605 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.621843 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52fe9600-233f-4c54-a6db-38d2a5e62d37" path="/var/lib/kubelet/pods/52fe9600-233f-4c54-a6db-38d2a5e62d37/volumes" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.624217 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e72f6a7-027d-4327-b9eb-39e28a4ab2e4" path="/var/lib/kubelet/pods/7e72f6a7-027d-4327-b9eb-39e28a4ab2e4/volumes" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.652880 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57968d6747-922m8"] Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.705026 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp"] Dec 08 19:28:25 crc kubenswrapper[4706]: W1208 19:28:25.712309 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89bf1c31_1caf_4a23_87e5_bbdba71df427.slice/crio-67476cbca206ad57f82815ae4bc3980d7f36ccefbb9887587332b304db12e398 WatchSource:0}: Error finding container 67476cbca206ad57f82815ae4bc3980d7f36ccefbb9887587332b304db12e398: Status 404 returned error can't find the container with id 67476cbca206ad57f82815ae4bc3980d7f36ccefbb9887587332b304db12e398 Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.888044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57968d6747-922m8" event={"ID":"3caf0864-a852-4a45-b426-d32b49dd4aa7","Type":"ContainerStarted","Data":"c3d5b44559e6bc817255d27c1a2cf96dbcdc27ecb9bf4c7abf235a20eb0dbbb0"} Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.888960 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.889045 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57968d6747-922m8" event={"ID":"3caf0864-a852-4a45-b426-d32b49dd4aa7","Type":"ContainerStarted","Data":"c9dba774800e7dff4a5f8cde9b1bb5cd38593d27c4fb6afde1cca1a118017c28"} Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.889535 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" event={"ID":"89bf1c31-1caf-4a23-87e5-bbdba71df427","Type":"ContainerStarted","Data":"37b38f259ed0b6c710357cedd59a396a27289fcfd426c79ec60ad769b62cc0df"} Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.890081 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.890100 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" event={"ID":"89bf1c31-1caf-4a23-87e5-bbdba71df427","Type":"ContainerStarted","Data":"67476cbca206ad57f82815ae4bc3980d7f36ccefbb9887587332b304db12e398"} Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.890776 4706 patch_prober.go:28] interesting pod/controller-manager-57968d6747-922m8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.890917 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57968d6747-922m8" podUID="3caf0864-a852-4a45-b426-d32b49dd4aa7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.891091 4706 patch_prober.go:28] interesting pod/route-controller-manager-76c99cf7b7-qcllp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.891139 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" podUID="89bf1c31-1caf-4a23-87e5-bbdba71df427" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.911693 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57968d6747-922m8" podStartSLOduration=2.911673952 podStartE2EDuration="2.911673952s" podCreationTimestamp="2025-12-08 19:28:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:28:25.910189258 +0000 UTC m=+388.552390261" watchObservedRunningTime="2025-12-08 19:28:25.911673952 +0000 UTC m=+388.553874955" Dec 08 19:28:25 crc kubenswrapper[4706]: I1208 19:28:25.930404 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" podStartSLOduration=2.9303790530000002 podStartE2EDuration="2.930379053s" podCreationTimestamp="2025-12-08 19:28:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:28:25.927040172 +0000 UTC m=+388.569241175" watchObservedRunningTime="2025-12-08 19:28:25.930379053 +0000 UTC m=+388.572580056" Dec 08 19:28:26 crc kubenswrapper[4706]: I1208 19:28:26.899024 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57968d6747-922m8" Dec 08 19:28:26 crc kubenswrapper[4706]: I1208 19:28:26.900582 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-qcllp" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.039608 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" podUID="25a09c7e-68bf-4a29-9bd5-909468bdcc42" containerName="registry" containerID="cri-o://422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990" gracePeriod=30 Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.511052 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.606741 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.606804 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/25a09c7e-68bf-4a29-9bd5-909468bdcc42-ca-trust-extracted\") pod \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.606832 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-tls\") pod \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.606861 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-trusted-ca\") pod \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.606936 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-bound-sa-token\") pod \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.607013 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/25a09c7e-68bf-4a29-9bd5-909468bdcc42-installation-pull-secrets\") pod \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.607059 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-certificates\") pod \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.607092 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2p9h\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-kube-api-access-l2p9h\") pod \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\" (UID: \"25a09c7e-68bf-4a29-9bd5-909468bdcc42\") " Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.608207 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "25a09c7e-68bf-4a29-9bd5-909468bdcc42" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.608520 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "25a09c7e-68bf-4a29-9bd5-909468bdcc42" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.609431 4706 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.609511 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25a09c7e-68bf-4a29-9bd5-909468bdcc42-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.619076 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a09c7e-68bf-4a29-9bd5-909468bdcc42-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "25a09c7e-68bf-4a29-9bd5-909468bdcc42" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.626658 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "25a09c7e-68bf-4a29-9bd5-909468bdcc42" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.634884 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-kube-api-access-l2p9h" (OuterVolumeSpecName: "kube-api-access-l2p9h") pod "25a09c7e-68bf-4a29-9bd5-909468bdcc42" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42"). InnerVolumeSpecName "kube-api-access-l2p9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.636096 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "25a09c7e-68bf-4a29-9bd5-909468bdcc42" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.641707 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25a09c7e-68bf-4a29-9bd5-909468bdcc42-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "25a09c7e-68bf-4a29-9bd5-909468bdcc42" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.642133 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "25a09c7e-68bf-4a29-9bd5-909468bdcc42" (UID: "25a09c7e-68bf-4a29-9bd5-909468bdcc42"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.710656 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.710989 4706 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/25a09c7e-68bf-4a29-9bd5-909468bdcc42-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.711009 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2p9h\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-kube-api-access-l2p9h\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.711022 4706 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/25a09c7e-68bf-4a29-9bd5-909468bdcc42-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.711036 4706 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/25a09c7e-68bf-4a29-9bd5-909468bdcc42-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.836470 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.836548 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.942972 4706 generic.go:334] "Generic (PLEG): container finished" podID="25a09c7e-68bf-4a29-9bd5-909468bdcc42" containerID="422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990" exitCode=0 Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.943049 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" event={"ID":"25a09c7e-68bf-4a29-9bd5-909468bdcc42","Type":"ContainerDied","Data":"422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990"} Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.943099 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.943178 4706 scope.go:117] "RemoveContainer" containerID="422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.943162 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-28lq2" event={"ID":"25a09c7e-68bf-4a29-9bd5-909468bdcc42","Type":"ContainerDied","Data":"13266bccd0bad2d5a1c3694b87553b9ee81bd8213ce29b5123df49a387374ed3"} Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.981693 4706 scope.go:117] "RemoveContainer" containerID="422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990" Dec 08 19:28:35 crc kubenswrapper[4706]: E1208 19:28:35.983600 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990\": container with ID starting with 422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990 not found: ID does not exist" containerID="422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.983641 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990"} err="failed to get container status \"422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990\": rpc error: code = NotFound desc = could not find container \"422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990\": container with ID starting with 422c4cd28368ebe60dd95e28fd1dc4825367040b516771cd76a466ac29114990 not found: ID does not exist" Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.987354 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-28lq2"] Dec 08 19:28:35 crc kubenswrapper[4706]: I1208 19:28:35.991174 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-28lq2"] Dec 08 19:28:37 crc kubenswrapper[4706]: I1208 19:28:37.616000 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25a09c7e-68bf-4a29-9bd5-909468bdcc42" path="/var/lib/kubelet/pods/25a09c7e-68bf-4a29-9bd5-909468bdcc42/volumes" Dec 08 19:29:05 crc kubenswrapper[4706]: I1208 19:29:05.836326 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:29:05 crc kubenswrapper[4706]: I1208 19:29:05.837084 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:29:05 crc kubenswrapper[4706]: I1208 19:29:05.837181 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:29:05 crc kubenswrapper[4706]: I1208 19:29:05.838004 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ab30946f44cee272e62060d2c6e0df51babab84dbf0a98d8b03bd4d916925ee"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:29:05 crc kubenswrapper[4706]: I1208 19:29:05.838075 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://9ab30946f44cee272e62060d2c6e0df51babab84dbf0a98d8b03bd4d916925ee" gracePeriod=600 Dec 08 19:29:06 crc kubenswrapper[4706]: I1208 19:29:06.139778 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="9ab30946f44cee272e62060d2c6e0df51babab84dbf0a98d8b03bd4d916925ee" exitCode=0 Dec 08 19:29:06 crc kubenswrapper[4706]: I1208 19:29:06.139884 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"9ab30946f44cee272e62060d2c6e0df51babab84dbf0a98d8b03bd4d916925ee"} Dec 08 19:29:06 crc kubenswrapper[4706]: I1208 19:29:06.140359 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"00eeebf4bd5f58c323dc15ecfd6b82b6e5fadacd75c5f1d8bddcf05b29f53f47"} Dec 08 19:29:06 crc kubenswrapper[4706]: I1208 19:29:06.140384 4706 scope.go:117] "RemoveContainer" containerID="ef8ba11b61a69851c5ccbe80021ef7fd217ba375bdbfb2cdd6c3405ce2caed12" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.194058 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b"] Dec 08 19:30:00 crc kubenswrapper[4706]: E1208 19:30:00.194993 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25a09c7e-68bf-4a29-9bd5-909468bdcc42" containerName="registry" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.195010 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="25a09c7e-68bf-4a29-9bd5-909468bdcc42" containerName="registry" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.195104 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="25a09c7e-68bf-4a29-9bd5-909468bdcc42" containerName="registry" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.195568 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.198254 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.198305 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.211518 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b"] Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.353635 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgfss\" (UniqueName: \"kubernetes.io/projected/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-kube-api-access-hgfss\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.353872 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-secret-volume\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.353993 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-config-volume\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.455008 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-secret-volume\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.455094 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-config-volume\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.455141 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgfss\" (UniqueName: \"kubernetes.io/projected/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-kube-api-access-hgfss\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.456443 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-config-volume\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.462327 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-secret-volume\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.477736 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgfss\" (UniqueName: \"kubernetes.io/projected/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-kube-api-access-hgfss\") pod \"collect-profiles-29420370-clr2b\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.516554 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:00 crc kubenswrapper[4706]: I1208 19:30:00.908695 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b"] Dec 08 19:30:01 crc kubenswrapper[4706]: I1208 19:30:01.482053 4706 generic.go:334] "Generic (PLEG): container finished" podID="c86e8022-6d78-4826-aa6e-55fd4c38bf4f" containerID="6b32975cff580efe0937b8686c83949688035d63341c5fbfb7bb778ecd5debe0" exitCode=0 Dec 08 19:30:01 crc kubenswrapper[4706]: I1208 19:30:01.482131 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" event={"ID":"c86e8022-6d78-4826-aa6e-55fd4c38bf4f","Type":"ContainerDied","Data":"6b32975cff580efe0937b8686c83949688035d63341c5fbfb7bb778ecd5debe0"} Dec 08 19:30:01 crc kubenswrapper[4706]: I1208 19:30:01.482584 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" event={"ID":"c86e8022-6d78-4826-aa6e-55fd4c38bf4f","Type":"ContainerStarted","Data":"d969127307e98ef0a77dc83c3c1abb607e00cc9e505f527923d7fe48773e229a"} Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.710619 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.790068 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-config-volume\") pod \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.790188 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgfss\" (UniqueName: \"kubernetes.io/projected/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-kube-api-access-hgfss\") pod \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.790233 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-secret-volume\") pod \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\" (UID: \"c86e8022-6d78-4826-aa6e-55fd4c38bf4f\") " Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.791373 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-config-volume" (OuterVolumeSpecName: "config-volume") pod "c86e8022-6d78-4826-aa6e-55fd4c38bf4f" (UID: "c86e8022-6d78-4826-aa6e-55fd4c38bf4f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.797141 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c86e8022-6d78-4826-aa6e-55fd4c38bf4f" (UID: "c86e8022-6d78-4826-aa6e-55fd4c38bf4f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.797224 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-kube-api-access-hgfss" (OuterVolumeSpecName: "kube-api-access-hgfss") pod "c86e8022-6d78-4826-aa6e-55fd4c38bf4f" (UID: "c86e8022-6d78-4826-aa6e-55fd4c38bf4f"). InnerVolumeSpecName "kube-api-access-hgfss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.892077 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.892142 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgfss\" (UniqueName: \"kubernetes.io/projected/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-kube-api-access-hgfss\") on node \"crc\" DevicePath \"\"" Dec 08 19:30:02 crc kubenswrapper[4706]: I1208 19:30:02.892159 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c86e8022-6d78-4826-aa6e-55fd4c38bf4f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 19:30:03 crc kubenswrapper[4706]: I1208 19:30:03.493070 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" event={"ID":"c86e8022-6d78-4826-aa6e-55fd4c38bf4f","Type":"ContainerDied","Data":"d969127307e98ef0a77dc83c3c1abb607e00cc9e505f527923d7fe48773e229a"} Dec 08 19:30:03 crc kubenswrapper[4706]: I1208 19:30:03.493111 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d969127307e98ef0a77dc83c3c1abb607e00cc9e505f527923d7fe48773e229a" Dec 08 19:30:03 crc kubenswrapper[4706]: I1208 19:30:03.493166 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b" Dec 08 19:31:35 crc kubenswrapper[4706]: I1208 19:31:35.835949 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:31:35 crc kubenswrapper[4706]: I1208 19:31:35.836864 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:31:57 crc kubenswrapper[4706]: I1208 19:31:57.854971 4706 scope.go:117] "RemoveContainer" containerID="71a60de3f04c9a2613bd658290cb3f010b3e0c44d5c2fc77fe444660a8c05d12" Dec 08 19:32:05 crc kubenswrapper[4706]: I1208 19:32:05.836386 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:32:05 crc kubenswrapper[4706]: I1208 19:32:05.837375 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:32:27 crc kubenswrapper[4706]: I1208 19:32:27.806091 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24"] Dec 08 19:32:27 crc kubenswrapper[4706]: E1208 19:32:27.807144 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86e8022-6d78-4826-aa6e-55fd4c38bf4f" containerName="collect-profiles" Dec 08 19:32:27 crc kubenswrapper[4706]: I1208 19:32:27.807158 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86e8022-6d78-4826-aa6e-55fd4c38bf4f" containerName="collect-profiles" Dec 08 19:32:27 crc kubenswrapper[4706]: I1208 19:32:27.807311 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c86e8022-6d78-4826-aa6e-55fd4c38bf4f" containerName="collect-profiles" Dec 08 19:32:27 crc kubenswrapper[4706]: I1208 19:32:27.808114 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:27 crc kubenswrapper[4706]: I1208 19:32:27.810220 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 08 19:32:27 crc kubenswrapper[4706]: I1208 19:32:27.824365 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24"] Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.001770 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w482h\" (UniqueName: \"kubernetes.io/projected/9a62102a-4f69-4782-9153-f2b7a3136127-kube-api-access-w482h\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.001833 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.001865 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.103141 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w482h\" (UniqueName: \"kubernetes.io/projected/9a62102a-4f69-4782-9153-f2b7a3136127-kube-api-access-w482h\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.103197 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.103222 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.103902 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.103970 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.124389 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w482h\" (UniqueName: \"kubernetes.io/projected/9a62102a-4f69-4782-9153-f2b7a3136127-kube-api-access-w482h\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.125821 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.328165 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24"] Dec 08 19:32:28 crc kubenswrapper[4706]: I1208 19:32:28.476785 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" event={"ID":"9a62102a-4f69-4782-9153-f2b7a3136127","Type":"ContainerStarted","Data":"491ad24cd15467ca9429ff59ecb84a734c86e7d520a59cb921d1f87feea77138"} Dec 08 19:32:29 crc kubenswrapper[4706]: I1208 19:32:29.486252 4706 generic.go:334] "Generic (PLEG): container finished" podID="9a62102a-4f69-4782-9153-f2b7a3136127" containerID="bd18eedc768980a24baace39859fd88a6314aadf7edfeb1ac6404a48eaead60f" exitCode=0 Dec 08 19:32:29 crc kubenswrapper[4706]: I1208 19:32:29.486330 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" event={"ID":"9a62102a-4f69-4782-9153-f2b7a3136127","Type":"ContainerDied","Data":"bd18eedc768980a24baace39859fd88a6314aadf7edfeb1ac6404a48eaead60f"} Dec 08 19:32:29 crc kubenswrapper[4706]: I1208 19:32:29.489226 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 19:32:31 crc kubenswrapper[4706]: I1208 19:32:31.499115 4706 generic.go:334] "Generic (PLEG): container finished" podID="9a62102a-4f69-4782-9153-f2b7a3136127" containerID="d28393941b991c3c6e116a7eb4d5b5e6554e3f3b2abd7d3f3b49a00d286ba9a2" exitCode=0 Dec 08 19:32:31 crc kubenswrapper[4706]: I1208 19:32:31.499551 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" event={"ID":"9a62102a-4f69-4782-9153-f2b7a3136127","Type":"ContainerDied","Data":"d28393941b991c3c6e116a7eb4d5b5e6554e3f3b2abd7d3f3b49a00d286ba9a2"} Dec 08 19:32:32 crc kubenswrapper[4706]: I1208 19:32:32.506512 4706 generic.go:334] "Generic (PLEG): container finished" podID="9a62102a-4f69-4782-9153-f2b7a3136127" containerID="d653f26d43b9dcc79d7b491cb00ee8086c258910a8507f4db17cbcf8fb4b7203" exitCode=0 Dec 08 19:32:32 crc kubenswrapper[4706]: I1208 19:32:32.506846 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" event={"ID":"9a62102a-4f69-4782-9153-f2b7a3136127","Type":"ContainerDied","Data":"d653f26d43b9dcc79d7b491cb00ee8086c258910a8507f4db17cbcf8fb4b7203"} Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.743283 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.883358 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-util\") pod \"9a62102a-4f69-4782-9153-f2b7a3136127\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.883453 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w482h\" (UniqueName: \"kubernetes.io/projected/9a62102a-4f69-4782-9153-f2b7a3136127-kube-api-access-w482h\") pod \"9a62102a-4f69-4782-9153-f2b7a3136127\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.883557 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-bundle\") pod \"9a62102a-4f69-4782-9153-f2b7a3136127\" (UID: \"9a62102a-4f69-4782-9153-f2b7a3136127\") " Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.886348 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-bundle" (OuterVolumeSpecName: "bundle") pod "9a62102a-4f69-4782-9153-f2b7a3136127" (UID: "9a62102a-4f69-4782-9153-f2b7a3136127"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.889643 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a62102a-4f69-4782-9153-f2b7a3136127-kube-api-access-w482h" (OuterVolumeSpecName: "kube-api-access-w482h") pod "9a62102a-4f69-4782-9153-f2b7a3136127" (UID: "9a62102a-4f69-4782-9153-f2b7a3136127"). InnerVolumeSpecName "kube-api-access-w482h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.897839 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-util" (OuterVolumeSpecName: "util") pod "9a62102a-4f69-4782-9153-f2b7a3136127" (UID: "9a62102a-4f69-4782-9153-f2b7a3136127"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.984668 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-util\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.984716 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w482h\" (UniqueName: \"kubernetes.io/projected/9a62102a-4f69-4782-9153-f2b7a3136127-kube-api-access-w482h\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:33 crc kubenswrapper[4706]: I1208 19:32:33.984727 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a62102a-4f69-4782-9153-f2b7a3136127-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:34 crc kubenswrapper[4706]: I1208 19:32:34.521414 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" event={"ID":"9a62102a-4f69-4782-9153-f2b7a3136127","Type":"ContainerDied","Data":"491ad24cd15467ca9429ff59ecb84a734c86e7d520a59cb921d1f87feea77138"} Dec 08 19:32:34 crc kubenswrapper[4706]: I1208 19:32:34.521510 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="491ad24cd15467ca9429ff59ecb84a734c86e7d520a59cb921d1f87feea77138" Dec 08 19:32:34 crc kubenswrapper[4706]: I1208 19:32:34.521457 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24" Dec 08 19:32:35 crc kubenswrapper[4706]: I1208 19:32:35.836723 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:32:35 crc kubenswrapper[4706]: I1208 19:32:35.837321 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:32:35 crc kubenswrapper[4706]: I1208 19:32:35.837400 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:32:35 crc kubenswrapper[4706]: I1208 19:32:35.838502 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"00eeebf4bd5f58c323dc15ecfd6b82b6e5fadacd75c5f1d8bddcf05b29f53f47"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:32:35 crc kubenswrapper[4706]: I1208 19:32:35.838650 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://00eeebf4bd5f58c323dc15ecfd6b82b6e5fadacd75c5f1d8bddcf05b29f53f47" gracePeriod=600 Dec 08 19:32:36 crc kubenswrapper[4706]: I1208 19:32:36.541855 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="00eeebf4bd5f58c323dc15ecfd6b82b6e5fadacd75c5f1d8bddcf05b29f53f47" exitCode=0 Dec 08 19:32:36 crc kubenswrapper[4706]: I1208 19:32:36.541923 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"00eeebf4bd5f58c323dc15ecfd6b82b6e5fadacd75c5f1d8bddcf05b29f53f47"} Dec 08 19:32:36 crc kubenswrapper[4706]: I1208 19:32:36.542001 4706 scope.go:117] "RemoveContainer" containerID="9ab30946f44cee272e62060d2c6e0df51babab84dbf0a98d8b03bd4d916925ee" Dec 08 19:32:37 crc kubenswrapper[4706]: I1208 19:32:37.554420 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"7d44a6096d27dcbd84b0ab9d9f054b00bc8d565b62f431d2f781917837a86f52"} Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.794945 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kl95h"] Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.795467 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovn-controller" containerID="cri-o://127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d" gracePeriod=30 Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.795573 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="nbdb" containerID="cri-o://a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62" gracePeriod=30 Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.795643 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142" gracePeriod=30 Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.795740 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="sbdb" containerID="cri-o://f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958" gracePeriod=30 Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.795722 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="northd" containerID="cri-o://3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819" gracePeriod=30 Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.795627 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kube-rbac-proxy-node" containerID="cri-o://b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1" gracePeriod=30 Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.795643 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovn-acl-logging" containerID="cri-o://1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e" gracePeriod=30 Dec 08 19:32:38 crc kubenswrapper[4706]: I1208 19:32:38.828072 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" containerID="cri-o://03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7" gracePeriod=30 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.180346 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/3.log" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.182677 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovn-acl-logging/0.log" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.183346 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovn-controller/0.log" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.184130 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241445 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-r4gr8"] Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241678 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241693 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241704 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241711 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241721 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a62102a-4f69-4782-9153-f2b7a3136127" containerName="util" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241729 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a62102a-4f69-4782-9153-f2b7a3136127" containerName="util" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241737 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241743 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241755 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovn-acl-logging" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241762 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovn-acl-logging" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241770 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kube-rbac-proxy-node" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241776 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kube-rbac-proxy-node" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241786 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="sbdb" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241791 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="sbdb" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241800 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a62102a-4f69-4782-9153-f2b7a3136127" containerName="pull" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241806 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a62102a-4f69-4782-9153-f2b7a3136127" containerName="pull" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241814 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="northd" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241821 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="northd" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241830 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kube-rbac-proxy-ovn-metrics" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241835 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kube-rbac-proxy-ovn-metrics" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241846 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kubecfg-setup" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241852 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kubecfg-setup" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241861 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="nbdb" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241866 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="nbdb" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241875 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a62102a-4f69-4782-9153-f2b7a3136127" containerName="extract" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241883 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a62102a-4f69-4782-9153-f2b7a3136127" containerName="extract" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.241891 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovn-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241897 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovn-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241985 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovn-acl-logging" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.241994 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242000 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="nbdb" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242009 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242016 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovn-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242026 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242033 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="northd" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242044 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kube-rbac-proxy-node" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242052 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="sbdb" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242060 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a62102a-4f69-4782-9153-f2b7a3136127" containerName="extract" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242067 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242076 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="kube-rbac-proxy-ovn-metrics" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.242163 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242170 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.242180 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242185 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.242337 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" containerName="ovnkube-controller" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.243988 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.256009 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a113dccf-6146-481e-947f-e782eb9c7b53-ovn-node-metrics-cert\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257305 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-openvswitch\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257381 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257476 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257556 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-node-log\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257509 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257688 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-node-log" (OuterVolumeSpecName: "node-log") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257642 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bptfb\" (UniqueName: \"kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257817 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-etc-openvswitch\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257895 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-netd\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.257972 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-ovn\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258078 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-var-lib-openvswitch\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258162 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-slash\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258242 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-kubelet\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258425 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-ovn-kubernetes\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258522 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-config\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258602 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-systemd-units\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258677 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-systemd\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258744 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-log-socket\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258819 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-bin\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258179 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258197 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258214 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258243 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258275 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-slash" (OuterVolumeSpecName: "host-slash") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258297 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258495 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258870 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258869 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258902 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-log-socket" (OuterVolumeSpecName: "log-socket") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258902 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.258916 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-netns\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259014 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-script-lib\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259046 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-env-overrides\") pod \"a113dccf-6146-481e-947f-e782eb9c7b53\" (UID: \"a113dccf-6146-481e-947f-e782eb9c7b53\") " Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259185 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-slash\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259212 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-ovn\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259234 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-run-netns\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259253 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-cni-netd\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259352 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-log-socket\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259412 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-systemd-units\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259493 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-cni-bin\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259540 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovnkube-script-lib\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259598 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-systemd\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259609 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259625 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-var-lib-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259683 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovn-node-metrics-cert\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259763 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259794 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovnkube-config\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259817 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259842 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-etc-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259869 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk57x\" (UniqueName: \"kubernetes.io/projected/571a262c-b9f2-4bb8-93b7-b744963fca2a-kube-api-access-sk57x\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259896 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-kubelet\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.259986 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260051 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-env-overrides\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260116 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-node-log\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260215 4706 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260236 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260283 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260300 4706 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260314 4706 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260327 4706 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-node-log\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260366 4706 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260383 4706 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260395 4706 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260407 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260420 4706 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-slash\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260460 4706 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260472 4706 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260483 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a113dccf-6146-481e-947f-e782eb9c7b53-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260494 4706 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260505 4706 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-log-socket\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.260796 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.264545 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a113dccf-6146-481e-947f-e782eb9c7b53-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.267415 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb" (OuterVolumeSpecName: "kube-api-access-bptfb") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "kube-api-access-bptfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.284018 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a113dccf-6146-481e-947f-e782eb9c7b53" (UID: "a113dccf-6146-481e-947f-e782eb9c7b53"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364641 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-systemd-units\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364743 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-cni-bin\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364781 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovnkube-script-lib\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364804 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-systemd\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364821 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-var-lib-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364840 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovn-node-metrics-cert\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364866 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364887 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovnkube-config\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364904 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364920 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-etc-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364938 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk57x\" (UniqueName: \"kubernetes.io/projected/571a262c-b9f2-4bb8-93b7-b744963fca2a-kube-api-access-sk57x\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364954 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-kubelet\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364972 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-env-overrides\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.364997 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-node-log\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365014 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-slash\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365033 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-ovn\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365048 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-cni-netd\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365063 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-run-netns\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365077 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365098 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-log-socket\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365156 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bptfb\" (UniqueName: \"kubernetes.io/projected/a113dccf-6146-481e-947f-e782eb9c7b53-kube-api-access-bptfb\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365168 4706 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365178 4706 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a113dccf-6146-481e-947f-e782eb9c7b53-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365188 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a113dccf-6146-481e-947f-e782eb9c7b53-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365233 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-log-socket\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365314 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-systemd-units\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.365340 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-cni-bin\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366318 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-systemd\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366376 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovnkube-script-lib\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366400 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-kubelet\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366568 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-var-lib-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366748 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-env-overrides\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366802 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-cni-netd\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366828 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-run-netns\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366852 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366879 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366906 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-etc-openvswitch\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366979 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-slash\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366981 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.366864 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-run-ovn\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.367032 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/571a262c-b9f2-4bb8-93b7-b744963fca2a-node-log\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.370405 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovnkube-config\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.372334 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/571a262c-b9f2-4bb8-93b7-b744963fca2a-ovn-node-metrics-cert\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.385242 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk57x\" (UniqueName: \"kubernetes.io/projected/571a262c-b9f2-4bb8-93b7-b744963fca2a-kube-api-access-sk57x\") pod \"ovnkube-node-r4gr8\" (UID: \"571a262c-b9f2-4bb8-93b7-b744963fca2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.557705 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.568708 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/2.log" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.569399 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/1.log" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.569475 4706 generic.go:334] "Generic (PLEG): container finished" podID="31d41aa9-3a74-42f4-9933-3f484a89c41d" containerID="17466bfe4afca9ffa7ad363b6a0c5b80401ec79daedb1e59ad77a6f7bc754587" exitCode=2 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.569552 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x674x" event={"ID":"31d41aa9-3a74-42f4-9933-3f484a89c41d","Type":"ContainerDied","Data":"17466bfe4afca9ffa7ad363b6a0c5b80401ec79daedb1e59ad77a6f7bc754587"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.569620 4706 scope.go:117] "RemoveContainer" containerID="f625c9701fbaba18236637655c6023db3da4523551465640fa68ff55ee904e22" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.570514 4706 scope.go:117] "RemoveContainer" containerID="17466bfe4afca9ffa7ad363b6a0c5b80401ec79daedb1e59ad77a6f7bc754587" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.570799 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x674x_openshift-multus(31d41aa9-3a74-42f4-9933-3f484a89c41d)\"" pod="openshift-multus/multus-x674x" podUID="31d41aa9-3a74-42f4-9933-3f484a89c41d" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.573230 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovnkube-controller/3.log" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.584945 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovn-acl-logging/0.log" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585493 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kl95h_a113dccf-6146-481e-947f-e782eb9c7b53/ovn-controller/0.log" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585837 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7" exitCode=0 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585864 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958" exitCode=0 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585874 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62" exitCode=0 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585884 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819" exitCode=0 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585891 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142" exitCode=0 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585898 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1" exitCode=0 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585906 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e" exitCode=143 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585914 4706 generic.go:334] "Generic (PLEG): container finished" podID="a113dccf-6146-481e-947f-e782eb9c7b53" containerID="127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d" exitCode=143 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585938 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585973 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.585992 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586009 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586020 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586030 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586045 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586057 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586065 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586071 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586077 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586082 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586088 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586094 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586099 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586104 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586112 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586120 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586126 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586131 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586136 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586141 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586146 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586152 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586158 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586164 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586169 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586176 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586183 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586189 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586195 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586201 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586207 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586212 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586218 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586224 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586231 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586237 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586245 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" event={"ID":"a113dccf-6146-481e-947f-e782eb9c7b53","Type":"ContainerDied","Data":"6777c21a19e6519123cc4ee0a68eaa48655bfc2dc2ad40cca941fbc6ae39ad6c"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586253 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586263 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586284 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586291 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586298 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586303 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586310 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586316 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586323 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586328 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.586483 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kl95h" Dec 08 19:32:39 crc kubenswrapper[4706]: W1208 19:32:39.602166 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod571a262c_b9f2_4bb8_93b7_b744963fca2a.slice/crio-5c8e80b0d0b8ffe547130e6a292336e8a3865c8de58a17477b9fbfef560c7296 WatchSource:0}: Error finding container 5c8e80b0d0b8ffe547130e6a292336e8a3865c8de58a17477b9fbfef560c7296: Status 404 returned error can't find the container with id 5c8e80b0d0b8ffe547130e6a292336e8a3865c8de58a17477b9fbfef560c7296 Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.624573 4706 scope.go:117] "RemoveContainer" containerID="03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.635164 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kl95h"] Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.641884 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kl95h"] Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.663636 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.698781 4706 scope.go:117] "RemoveContainer" containerID="f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.716445 4706 scope.go:117] "RemoveContainer" containerID="a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.735443 4706 scope.go:117] "RemoveContainer" containerID="3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.751989 4706 scope.go:117] "RemoveContainer" containerID="043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.769014 4706 scope.go:117] "RemoveContainer" containerID="b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.785367 4706 scope.go:117] "RemoveContainer" containerID="1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.802843 4706 scope.go:117] "RemoveContainer" containerID="127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.826503 4706 scope.go:117] "RemoveContainer" containerID="f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.899491 4706 scope.go:117] "RemoveContainer" containerID="03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.903374 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": container with ID starting with 03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7 not found: ID does not exist" containerID="03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.903418 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} err="failed to get container status \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": rpc error: code = NotFound desc = could not find container \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": container with ID starting with 03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.903452 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.903806 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": container with ID starting with e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4 not found: ID does not exist" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.903861 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} err="failed to get container status \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": rpc error: code = NotFound desc = could not find container \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": container with ID starting with e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.903893 4706 scope.go:117] "RemoveContainer" containerID="f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.904339 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": container with ID starting with f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958 not found: ID does not exist" containerID="f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.904373 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} err="failed to get container status \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": rpc error: code = NotFound desc = could not find container \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": container with ID starting with f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.904389 4706 scope.go:117] "RemoveContainer" containerID="a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.904695 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": container with ID starting with a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62 not found: ID does not exist" containerID="a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.904832 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} err="failed to get container status \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": rpc error: code = NotFound desc = could not find container \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": container with ID starting with a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.904947 4706 scope.go:117] "RemoveContainer" containerID="3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.906683 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": container with ID starting with 3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819 not found: ID does not exist" containerID="3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.906781 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} err="failed to get container status \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": rpc error: code = NotFound desc = could not find container \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": container with ID starting with 3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.906850 4706 scope.go:117] "RemoveContainer" containerID="043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.907240 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": container with ID starting with 043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142 not found: ID does not exist" containerID="043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.907336 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} err="failed to get container status \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": rpc error: code = NotFound desc = could not find container \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": container with ID starting with 043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.907402 4706 scope.go:117] "RemoveContainer" containerID="b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.907706 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": container with ID starting with b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1 not found: ID does not exist" containerID="b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.907779 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} err="failed to get container status \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": rpc error: code = NotFound desc = could not find container \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": container with ID starting with b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.907839 4706 scope.go:117] "RemoveContainer" containerID="1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.908090 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": container with ID starting with 1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e not found: ID does not exist" containerID="1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.908166 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} err="failed to get container status \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": rpc error: code = NotFound desc = could not find container \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": container with ID starting with 1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.908245 4706 scope.go:117] "RemoveContainer" containerID="127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.908550 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": container with ID starting with 127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d not found: ID does not exist" containerID="127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.908641 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} err="failed to get container status \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": rpc error: code = NotFound desc = could not find container \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": container with ID starting with 127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.908707 4706 scope.go:117] "RemoveContainer" containerID="f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944" Dec 08 19:32:39 crc kubenswrapper[4706]: E1208 19:32:39.912444 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": container with ID starting with f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944 not found: ID does not exist" containerID="f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.912592 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} err="failed to get container status \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": rpc error: code = NotFound desc = could not find container \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": container with ID starting with f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.912673 4706 scope.go:117] "RemoveContainer" containerID="03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.913034 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} err="failed to get container status \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": rpc error: code = NotFound desc = could not find container \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": container with ID starting with 03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.913124 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.913443 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} err="failed to get container status \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": rpc error: code = NotFound desc = could not find container \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": container with ID starting with e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.913532 4706 scope.go:117] "RemoveContainer" containerID="f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.913803 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} err="failed to get container status \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": rpc error: code = NotFound desc = could not find container \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": container with ID starting with f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.913882 4706 scope.go:117] "RemoveContainer" containerID="a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.914146 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} err="failed to get container status \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": rpc error: code = NotFound desc = could not find container \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": container with ID starting with a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.914244 4706 scope.go:117] "RemoveContainer" containerID="3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.915058 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} err="failed to get container status \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": rpc error: code = NotFound desc = could not find container \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": container with ID starting with 3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.915138 4706 scope.go:117] "RemoveContainer" containerID="043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.915426 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} err="failed to get container status \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": rpc error: code = NotFound desc = could not find container \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": container with ID starting with 043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.915508 4706 scope.go:117] "RemoveContainer" containerID="b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.922006 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} err="failed to get container status \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": rpc error: code = NotFound desc = could not find container \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": container with ID starting with b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.922178 4706 scope.go:117] "RemoveContainer" containerID="1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.926062 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} err="failed to get container status \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": rpc error: code = NotFound desc = could not find container \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": container with ID starting with 1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.926167 4706 scope.go:117] "RemoveContainer" containerID="127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.926514 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} err="failed to get container status \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": rpc error: code = NotFound desc = could not find container \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": container with ID starting with 127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.926589 4706 scope.go:117] "RemoveContainer" containerID="f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.927255 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} err="failed to get container status \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": rpc error: code = NotFound desc = could not find container \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": container with ID starting with f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.927415 4706 scope.go:117] "RemoveContainer" containerID="03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.927726 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} err="failed to get container status \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": rpc error: code = NotFound desc = could not find container \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": container with ID starting with 03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.927822 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.928081 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} err="failed to get container status \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": rpc error: code = NotFound desc = could not find container \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": container with ID starting with e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.928155 4706 scope.go:117] "RemoveContainer" containerID="f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.928418 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} err="failed to get container status \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": rpc error: code = NotFound desc = could not find container \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": container with ID starting with f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.928489 4706 scope.go:117] "RemoveContainer" containerID="a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.928743 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} err="failed to get container status \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": rpc error: code = NotFound desc = could not find container \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": container with ID starting with a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.928811 4706 scope.go:117] "RemoveContainer" containerID="3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.929048 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} err="failed to get container status \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": rpc error: code = NotFound desc = could not find container \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": container with ID starting with 3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.929115 4706 scope.go:117] "RemoveContainer" containerID="043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.931284 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} err="failed to get container status \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": rpc error: code = NotFound desc = could not find container \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": container with ID starting with 043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.931400 4706 scope.go:117] "RemoveContainer" containerID="b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.932729 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} err="failed to get container status \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": rpc error: code = NotFound desc = could not find container \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": container with ID starting with b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.932813 4706 scope.go:117] "RemoveContainer" containerID="1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.934970 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} err="failed to get container status \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": rpc error: code = NotFound desc = could not find container \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": container with ID starting with 1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.935071 4706 scope.go:117] "RemoveContainer" containerID="127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.935493 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} err="failed to get container status \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": rpc error: code = NotFound desc = could not find container \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": container with ID starting with 127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.935571 4706 scope.go:117] "RemoveContainer" containerID="f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.937094 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} err="failed to get container status \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": rpc error: code = NotFound desc = could not find container \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": container with ID starting with f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.937178 4706 scope.go:117] "RemoveContainer" containerID="03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.939948 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7"} err="failed to get container status \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": rpc error: code = NotFound desc = could not find container \"03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7\": container with ID starting with 03231b4a3ceaf5ee4352ab675207115290a7a2241f05561cb1a5e2ad9405fec7 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.940033 4706 scope.go:117] "RemoveContainer" containerID="e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.940373 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4"} err="failed to get container status \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": rpc error: code = NotFound desc = could not find container \"e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4\": container with ID starting with e29dd42bfdc8707c455b745c61b12b06b8209426cd8bd49131310ef049b381b4 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.940458 4706 scope.go:117] "RemoveContainer" containerID="f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.940726 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958"} err="failed to get container status \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": rpc error: code = NotFound desc = could not find container \"f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958\": container with ID starting with f60d60126b3c951c11f1f9d4b8010bb69e0942b0ce87d931484389829d75d958 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.940795 4706 scope.go:117] "RemoveContainer" containerID="a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.941448 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62"} err="failed to get container status \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": rpc error: code = NotFound desc = could not find container \"a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62\": container with ID starting with a937ad8f9f61a39ece6aad81e04e7a62ae8bba040855052f52ba929f7d4f5e62 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.941524 4706 scope.go:117] "RemoveContainer" containerID="3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.942241 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819"} err="failed to get container status \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": rpc error: code = NotFound desc = could not find container \"3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819\": container with ID starting with 3c9a6a9b3b7898fda732c429302de43b868ae3d92771e6d754d16ba7ddf91819 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.942359 4706 scope.go:117] "RemoveContainer" containerID="043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.942652 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142"} err="failed to get container status \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": rpc error: code = NotFound desc = could not find container \"043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142\": container with ID starting with 043e4c84bae835f9ab3cc050f32bab226cf4e893d9c8d61801e2dba64e00d142 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.942741 4706 scope.go:117] "RemoveContainer" containerID="b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.943041 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1"} err="failed to get container status \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": rpc error: code = NotFound desc = could not find container \"b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1\": container with ID starting with b2bf88f3b2578bcb815d46b898e10741b9d5aae344f9b1f93385e2d2dae785e1 not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.943133 4706 scope.go:117] "RemoveContainer" containerID="1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.943728 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e"} err="failed to get container status \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": rpc error: code = NotFound desc = could not find container \"1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e\": container with ID starting with 1929b94259f48243fa3340ffaa58714ee7ac82f28c9f7f7bab6196e11ed4000e not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.943819 4706 scope.go:117] "RemoveContainer" containerID="127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.944157 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d"} err="failed to get container status \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": rpc error: code = NotFound desc = could not find container \"127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d\": container with ID starting with 127474f3bb57abd92f62222f56867388f9adf3eb3985b6ec2729b124c90ae32d not found: ID does not exist" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.944249 4706 scope.go:117] "RemoveContainer" containerID="f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944" Dec 08 19:32:39 crc kubenswrapper[4706]: I1208 19:32:39.946808 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944"} err="failed to get container status \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": rpc error: code = NotFound desc = could not find container \"f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944\": container with ID starting with f7cfd2a9e8cb19d6f56fe8a8cfa68a8ef4a4921adb5b02f3812ebb62f4e86944 not found: ID does not exist" Dec 08 19:32:40 crc kubenswrapper[4706]: I1208 19:32:40.593529 4706 generic.go:334] "Generic (PLEG): container finished" podID="571a262c-b9f2-4bb8-93b7-b744963fca2a" containerID="ece3784cb67223a0343922eac92aeab1b20b66e6cb56f6014592889f350aacc4" exitCode=0 Dec 08 19:32:40 crc kubenswrapper[4706]: I1208 19:32:40.593766 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerDied","Data":"ece3784cb67223a0343922eac92aeab1b20b66e6cb56f6014592889f350aacc4"} Dec 08 19:32:40 crc kubenswrapper[4706]: I1208 19:32:40.593847 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"5c8e80b0d0b8ffe547130e6a292336e8a3865c8de58a17477b9fbfef560c7296"} Dec 08 19:32:40 crc kubenswrapper[4706]: I1208 19:32:40.597439 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/2.log" Dec 08 19:32:41 crc kubenswrapper[4706]: I1208 19:32:41.613792 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a113dccf-6146-481e-947f-e782eb9c7b53" path="/var/lib/kubelet/pods/a113dccf-6146-481e-947f-e782eb9c7b53/volumes" Dec 08 19:32:41 crc kubenswrapper[4706]: I1208 19:32:41.616518 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"efd9e0aa353441ffa7db92d4841f34b21d3c36e5b049aa9d45c5a6b75d1d4e8e"} Dec 08 19:32:41 crc kubenswrapper[4706]: I1208 19:32:41.616604 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"2bbd22ce5f2ae3872b2cbaabf0ffe13f838d4aedc8f295ba8990d36c4fc125c8"} Dec 08 19:32:41 crc kubenswrapper[4706]: I1208 19:32:41.616660 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"979ae7938d1fe15be3cb68151d92b8fab1b559ed9893ef6de6bbef2d3023f2e7"} Dec 08 19:32:41 crc kubenswrapper[4706]: I1208 19:32:41.616720 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"f622b1208c7be714e00c7ab42c1267cb18fadf83ed29fb9743457c2c4d5fc124"} Dec 08 19:32:41 crc kubenswrapper[4706]: I1208 19:32:41.616773 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"b416dcf5953ba6d65137429b7606f1865846a7c95dde4ebc86a5eae798178c04"} Dec 08 19:32:41 crc kubenswrapper[4706]: I1208 19:32:41.616831 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"232b83a3f6aac2e3dad7375e407a1e56dbb5861b75284c9953f195f9358d224e"} Dec 08 19:32:44 crc kubenswrapper[4706]: I1208 19:32:44.640854 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"65b73ef526428bdf628668b22d9de5272ac5b5e8e1a432a2830ae3af60c3bb40"} Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.389356 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr"] Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.391058 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.395868 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.395975 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-jrt5h" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.396085 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.439434 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx"] Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.440412 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.442879 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-jlnh6" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.442913 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.449786 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp"] Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.450537 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.568733 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f543c2b7-0b00-458c-b4e5-a8c89cfaff5c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp\" (UID: \"f543c2b7-0b00-458c-b4e5-a8c89cfaff5c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.568835 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f251b1b-6576-4421-aa76-04a26d0e1302-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx\" (UID: \"3f251b1b-6576-4421-aa76-04a26d0e1302\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.568915 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f251b1b-6576-4421-aa76-04a26d0e1302-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx\" (UID: \"3f251b1b-6576-4421-aa76-04a26d0e1302\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.568953 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f543c2b7-0b00-458c-b4e5-a8c89cfaff5c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp\" (UID: \"f543c2b7-0b00-458c-b4e5-a8c89cfaff5c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.569019 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmg6f\" (UniqueName: \"kubernetes.io/projected/90a15b07-3ae8-4265-a92b-d9dd052211e2-kube-api-access-vmg6f\") pod \"obo-prometheus-operator-668cf9dfbb-fntdr\" (UID: \"90a15b07-3ae8-4265-a92b-d9dd052211e2\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.648965 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-4xlqn"] Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.649907 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.652226 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.655730 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-tqr6q" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.656031 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" event={"ID":"571a262c-b9f2-4bb8-93b7-b744963fca2a","Type":"ContainerStarted","Data":"36cb1b2f6738aac5b4d9f17f92c68738d8dec597e0f44aac276a9e56bc8a8cf4"} Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.656426 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.656474 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.656640 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.670872 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f543c2b7-0b00-458c-b4e5-a8c89cfaff5c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp\" (UID: \"f543c2b7-0b00-458c-b4e5-a8c89cfaff5c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.670935 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmg6f\" (UniqueName: \"kubernetes.io/projected/90a15b07-3ae8-4265-a92b-d9dd052211e2-kube-api-access-vmg6f\") pod \"obo-prometheus-operator-668cf9dfbb-fntdr\" (UID: \"90a15b07-3ae8-4265-a92b-d9dd052211e2\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.670972 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f543c2b7-0b00-458c-b4e5-a8c89cfaff5c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp\" (UID: \"f543c2b7-0b00-458c-b4e5-a8c89cfaff5c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.671015 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f251b1b-6576-4421-aa76-04a26d0e1302-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx\" (UID: \"3f251b1b-6576-4421-aa76-04a26d0e1302\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.671078 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f251b1b-6576-4421-aa76-04a26d0e1302-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx\" (UID: \"3f251b1b-6576-4421-aa76-04a26d0e1302\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.681096 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f543c2b7-0b00-458c-b4e5-a8c89cfaff5c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp\" (UID: \"f543c2b7-0b00-458c-b4e5-a8c89cfaff5c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.682405 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f543c2b7-0b00-458c-b4e5-a8c89cfaff5c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp\" (UID: \"f543c2b7-0b00-458c-b4e5-a8c89cfaff5c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.687083 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f251b1b-6576-4421-aa76-04a26d0e1302-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx\" (UID: \"3f251b1b-6576-4421-aa76-04a26d0e1302\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.694806 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f251b1b-6576-4421-aa76-04a26d0e1302-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx\" (UID: \"3f251b1b-6576-4421-aa76-04a26d0e1302\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.695722 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.704182 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.710824 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmg6f\" (UniqueName: \"kubernetes.io/projected/90a15b07-3ae8-4265-a92b-d9dd052211e2-kube-api-access-vmg6f\") pod \"obo-prometheus-operator-668cf9dfbb-fntdr\" (UID: \"90a15b07-3ae8-4265-a92b-d9dd052211e2\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.716590 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.760540 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" podStartSLOduration=7.760509428 podStartE2EDuration="7.760509428s" podCreationTimestamp="2025-12-08 19:32:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:32:46.734412274 +0000 UTC m=+649.376613297" watchObservedRunningTime="2025-12-08 19:32:46.760509428 +0000 UTC m=+649.402710431" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.753102 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(04ed56a0e8df068761662e335956be2fba3999480b53883e62963af63f1889d7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.761783 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(04ed56a0e8df068761662e335956be2fba3999480b53883e62963af63f1889d7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.761817 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(04ed56a0e8df068761662e335956be2fba3999480b53883e62963af63f1889d7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.761876 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators(90a15b07-3ae8-4265-a92b-d9dd052211e2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators(90a15b07-3ae8-4265-a92b-d9dd052211e2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(04ed56a0e8df068761662e335956be2fba3999480b53883e62963af63f1889d7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" podUID="90a15b07-3ae8-4265-a92b-d9dd052211e2" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.762181 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.772454 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3cbb9b0b-a2c7-47f8-92d2-de4829687a45-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-4xlqn\" (UID: \"3cbb9b0b-a2c7-47f8-92d2-de4829687a45\") " pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.772582 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvrpl\" (UniqueName: \"kubernetes.io/projected/3cbb9b0b-a2c7-47f8-92d2-de4829687a45-kube-api-access-jvrpl\") pod \"observability-operator-d8bb48f5d-4xlqn\" (UID: \"3cbb9b0b-a2c7-47f8-92d2-de4829687a45\") " pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.772757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.805102 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(d917a3b49237e4656b2321dd48f2c637ca4e968571ecd2fa10f99396c921d478): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.805205 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(d917a3b49237e4656b2321dd48f2c637ca4e968571ecd2fa10f99396c921d478): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.805234 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(d917a3b49237e4656b2321dd48f2c637ca4e968571ecd2fa10f99396c921d478): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.805334 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators(f543c2b7-0b00-458c-b4e5-a8c89cfaff5c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators(f543c2b7-0b00-458c-b4e5-a8c89cfaff5c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(d917a3b49237e4656b2321dd48f2c637ca4e968571ecd2fa10f99396c921d478): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" podUID="f543c2b7-0b00-458c-b4e5-a8c89cfaff5c" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.824108 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(9bf9c8b4a1e2064b779d67793a8bb82bf997e1d216dce6964dc2d04d03815fff): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.824191 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(9bf9c8b4a1e2064b779d67793a8bb82bf997e1d216dce6964dc2d04d03815fff): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.824218 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(9bf9c8b4a1e2064b779d67793a8bb82bf997e1d216dce6964dc2d04d03815fff): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:46 crc kubenswrapper[4706]: E1208 19:32:46.824308 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators(3f251b1b-6576-4421-aa76-04a26d0e1302)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators(3f251b1b-6576-4421-aa76-04a26d0e1302)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(9bf9c8b4a1e2064b779d67793a8bb82bf997e1d216dce6964dc2d04d03815fff): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" podUID="3f251b1b-6576-4421-aa76-04a26d0e1302" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.832037 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-fqdzc"] Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.832828 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.836768 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-nt2tq" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.874438 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvrpl\" (UniqueName: \"kubernetes.io/projected/3cbb9b0b-a2c7-47f8-92d2-de4829687a45-kube-api-access-jvrpl\") pod \"observability-operator-d8bb48f5d-4xlqn\" (UID: \"3cbb9b0b-a2c7-47f8-92d2-de4829687a45\") " pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.874729 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3cbb9b0b-a2c7-47f8-92d2-de4829687a45-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-4xlqn\" (UID: \"3cbb9b0b-a2c7-47f8-92d2-de4829687a45\") " pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.878679 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3cbb9b0b-a2c7-47f8-92d2-de4829687a45-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-4xlqn\" (UID: \"3cbb9b0b-a2c7-47f8-92d2-de4829687a45\") " pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.898186 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvrpl\" (UniqueName: \"kubernetes.io/projected/3cbb9b0b-a2c7-47f8-92d2-de4829687a45-kube-api-access-jvrpl\") pod \"observability-operator-d8bb48f5d-4xlqn\" (UID: \"3cbb9b0b-a2c7-47f8-92d2-de4829687a45\") " pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.975828 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.976191 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2dp8\" (UniqueName: \"kubernetes.io/projected/a8e2aac6-1870-4ddc-8650-c31ded55337f-kube-api-access-b2dp8\") pod \"perses-operator-5446b9c989-fqdzc\" (UID: \"a8e2aac6-1870-4ddc-8650-c31ded55337f\") " pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:46 crc kubenswrapper[4706]: I1208 19:32:46.976255 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a8e2aac6-1870-4ddc-8650-c31ded55337f-openshift-service-ca\") pod \"perses-operator-5446b9c989-fqdzc\" (UID: \"a8e2aac6-1870-4ddc-8650-c31ded55337f\") " pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.005141 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(95260e6574ab46c2f61b0f56b8fb6328cfed22d32a775a657df85f997dafea2e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.005221 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(95260e6574ab46c2f61b0f56b8fb6328cfed22d32a775a657df85f997dafea2e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.005246 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(95260e6574ab46c2f61b0f56b8fb6328cfed22d32a775a657df85f997dafea2e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.005379 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-4xlqn_openshift-operators(3cbb9b0b-a2c7-47f8-92d2-de4829687a45)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-4xlqn_openshift-operators(3cbb9b0b-a2c7-47f8-92d2-de4829687a45)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(95260e6574ab46c2f61b0f56b8fb6328cfed22d32a775a657df85f997dafea2e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" podUID="3cbb9b0b-a2c7-47f8-92d2-de4829687a45" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.077368 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2dp8\" (UniqueName: \"kubernetes.io/projected/a8e2aac6-1870-4ddc-8650-c31ded55337f-kube-api-access-b2dp8\") pod \"perses-operator-5446b9c989-fqdzc\" (UID: \"a8e2aac6-1870-4ddc-8650-c31ded55337f\") " pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.077429 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a8e2aac6-1870-4ddc-8650-c31ded55337f-openshift-service-ca\") pod \"perses-operator-5446b9c989-fqdzc\" (UID: \"a8e2aac6-1870-4ddc-8650-c31ded55337f\") " pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.078299 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a8e2aac6-1870-4ddc-8650-c31ded55337f-openshift-service-ca\") pod \"perses-operator-5446b9c989-fqdzc\" (UID: \"a8e2aac6-1870-4ddc-8650-c31ded55337f\") " pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.119125 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2dp8\" (UniqueName: \"kubernetes.io/projected/a8e2aac6-1870-4ddc-8650-c31ded55337f-kube-api-access-b2dp8\") pod \"perses-operator-5446b9c989-fqdzc\" (UID: \"a8e2aac6-1870-4ddc-8650-c31ded55337f\") " pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.154744 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.183866 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(a5e862b33085012fadbdaad9a0d769f09f65988f69b52e592dad4fffd0df5d80): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.183944 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(a5e862b33085012fadbdaad9a0d769f09f65988f69b52e592dad4fffd0df5d80): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.183966 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(a5e862b33085012fadbdaad9a0d769f09f65988f69b52e592dad4fffd0df5d80): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.184013 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-fqdzc_openshift-operators(a8e2aac6-1870-4ddc-8650-c31ded55337f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-fqdzc_openshift-operators(a8e2aac6-1870-4ddc-8650-c31ded55337f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(a5e862b33085012fadbdaad9a0d769f09f65988f69b52e592dad4fffd0df5d80): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" podUID="a8e2aac6-1870-4ddc-8650-c31ded55337f" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.427289 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp"] Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.453115 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr"] Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.458628 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-fqdzc"] Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.471505 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx"] Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.480906 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-4xlqn"] Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.661358 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.661410 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.661452 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.661473 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.661520 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.662486 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.662952 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.662972 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.663323 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:47 crc kubenswrapper[4706]: I1208 19:32:47.663547 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.715786 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(4cebd4ebe6927f1efb0577c3e66ac16453bff3260ba196c79024ef20edf21a4c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.715890 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(4cebd4ebe6927f1efb0577c3e66ac16453bff3260ba196c79024ef20edf21a4c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.715928 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(4cebd4ebe6927f1efb0577c3e66ac16453bff3260ba196c79024ef20edf21a4c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.715994 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators(90a15b07-3ae8-4265-a92b-d9dd052211e2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators(90a15b07-3ae8-4265-a92b-d9dd052211e2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(4cebd4ebe6927f1efb0577c3e66ac16453bff3260ba196c79024ef20edf21a4c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" podUID="90a15b07-3ae8-4265-a92b-d9dd052211e2" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.731120 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(f2505a99c3b9ba611c88afd3a701e73832571bc28f0ed0a502961a35ade1809d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.731232 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(f2505a99c3b9ba611c88afd3a701e73832571bc28f0ed0a502961a35ade1809d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.731276 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(f2505a99c3b9ba611c88afd3a701e73832571bc28f0ed0a502961a35ade1809d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.731358 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-4xlqn_openshift-operators(3cbb9b0b-a2c7-47f8-92d2-de4829687a45)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-4xlqn_openshift-operators(3cbb9b0b-a2c7-47f8-92d2-de4829687a45)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(f2505a99c3b9ba611c88afd3a701e73832571bc28f0ed0a502961a35ade1809d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" podUID="3cbb9b0b-a2c7-47f8-92d2-de4829687a45" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.738140 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(88c7da3dd2917eeba228b54626498b56fbfa952d548916024ae947b83af10a3c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.738248 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(88c7da3dd2917eeba228b54626498b56fbfa952d548916024ae947b83af10a3c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.738310 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(88c7da3dd2917eeba228b54626498b56fbfa952d548916024ae947b83af10a3c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.738379 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators(3f251b1b-6576-4421-aa76-04a26d0e1302)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators(3f251b1b-6576-4421-aa76-04a26d0e1302)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(88c7da3dd2917eeba228b54626498b56fbfa952d548916024ae947b83af10a3c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" podUID="3f251b1b-6576-4421-aa76-04a26d0e1302" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.773753 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(83ac423bdff2e1a0bd0c080e8386b006b31d044dd04df6ccf8acd2e950444d07): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.773843 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(83ac423bdff2e1a0bd0c080e8386b006b31d044dd04df6ccf8acd2e950444d07): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.773873 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(83ac423bdff2e1a0bd0c080e8386b006b31d044dd04df6ccf8acd2e950444d07): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.773923 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-fqdzc_openshift-operators(a8e2aac6-1870-4ddc-8650-c31ded55337f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-fqdzc_openshift-operators(a8e2aac6-1870-4ddc-8650-c31ded55337f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(83ac423bdff2e1a0bd0c080e8386b006b31d044dd04df6ccf8acd2e950444d07): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" podUID="a8e2aac6-1870-4ddc-8650-c31ded55337f" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.789613 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(a6735e6bf47577297e79e9b6842fef25bf1aeec9ace31e5566a75fc061213e69): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.789698 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(a6735e6bf47577297e79e9b6842fef25bf1aeec9ace31e5566a75fc061213e69): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.789724 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(a6735e6bf47577297e79e9b6842fef25bf1aeec9ace31e5566a75fc061213e69): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:32:47 crc kubenswrapper[4706]: E1208 19:32:47.789783 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators(f543c2b7-0b00-458c-b4e5-a8c89cfaff5c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators(f543c2b7-0b00-458c-b4e5-a8c89cfaff5c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(a6735e6bf47577297e79e9b6842fef25bf1aeec9ace31e5566a75fc061213e69): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" podUID="f543c2b7-0b00-458c-b4e5-a8c89cfaff5c" Dec 08 19:32:51 crc kubenswrapper[4706]: I1208 19:32:51.608360 4706 scope.go:117] "RemoveContainer" containerID="17466bfe4afca9ffa7ad363b6a0c5b80401ec79daedb1e59ad77a6f7bc754587" Dec 08 19:32:51 crc kubenswrapper[4706]: E1208 19:32:51.608931 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x674x_openshift-multus(31d41aa9-3a74-42f4-9933-3f484a89c41d)\"" pod="openshift-multus/multus-x674x" podUID="31d41aa9-3a74-42f4-9933-3f484a89c41d" Dec 08 19:32:59 crc kubenswrapper[4706]: I1208 19:32:59.608331 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:59 crc kubenswrapper[4706]: I1208 19:32:59.608372 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:59 crc kubenswrapper[4706]: I1208 19:32:59.608451 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:59 crc kubenswrapper[4706]: I1208 19:32:59.609493 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:59 crc kubenswrapper[4706]: I1208 19:32:59.609876 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:59 crc kubenswrapper[4706]: I1208 19:32:59.609876 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.666075 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(930e92d69d2face2ae8524a8da191f6275eab26697433fcc181d57809ee546de): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.666170 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(930e92d69d2face2ae8524a8da191f6275eab26697433fcc181d57809ee546de): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.666197 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(930e92d69d2face2ae8524a8da191f6275eab26697433fcc181d57809ee546de): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.666254 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators(3f251b1b-6576-4421-aa76-04a26d0e1302)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators(3f251b1b-6576-4421-aa76-04a26d0e1302)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_openshift-operators_3f251b1b-6576-4421-aa76-04a26d0e1302_0(930e92d69d2face2ae8524a8da191f6275eab26697433fcc181d57809ee546de): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" podUID="3f251b1b-6576-4421-aa76-04a26d0e1302" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.672202 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(4382ec4f40ec07001280a5be442a08a19dc7136b3cfca0b66f84f8fae051fba1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.672370 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(4382ec4f40ec07001280a5be442a08a19dc7136b3cfca0b66f84f8fae051fba1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.672406 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(4382ec4f40ec07001280a5be442a08a19dc7136b3cfca0b66f84f8fae051fba1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.672467 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-4xlqn_openshift-operators(3cbb9b0b-a2c7-47f8-92d2-de4829687a45)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-4xlqn_openshift-operators(3cbb9b0b-a2c7-47f8-92d2-de4829687a45)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-4xlqn_openshift-operators_3cbb9b0b-a2c7-47f8-92d2-de4829687a45_0(4382ec4f40ec07001280a5be442a08a19dc7136b3cfca0b66f84f8fae051fba1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" podUID="3cbb9b0b-a2c7-47f8-92d2-de4829687a45" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.676885 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(fa4ad473fc5c21e866653162f91fcabda7ceaf3e2c5fe0194c2f7363bf6a9d5f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.676948 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(fa4ad473fc5c21e866653162f91fcabda7ceaf3e2c5fe0194c2f7363bf6a9d5f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.676973 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(fa4ad473fc5c21e866653162f91fcabda7ceaf3e2c5fe0194c2f7363bf6a9d5f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:32:59 crc kubenswrapper[4706]: E1208 19:32:59.677030 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators(90a15b07-3ae8-4265-a92b-d9dd052211e2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators(90a15b07-3ae8-4265-a92b-d9dd052211e2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-fntdr_openshift-operators_90a15b07-3ae8-4265-a92b-d9dd052211e2_0(fa4ad473fc5c21e866653162f91fcabda7ceaf3e2c5fe0194c2f7363bf6a9d5f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" podUID="90a15b07-3ae8-4265-a92b-d9dd052211e2" Dec 08 19:33:00 crc kubenswrapper[4706]: I1208 19:33:00.607848 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:33:00 crc kubenswrapper[4706]: I1208 19:33:00.607995 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:33:00 crc kubenswrapper[4706]: I1208 19:33:00.608526 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:33:00 crc kubenswrapper[4706]: I1208 19:33:00.609043 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:33:00 crc kubenswrapper[4706]: E1208 19:33:00.638782 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(6fc26423a36b7251af2cd16592aedcd9217b851e8f334dcecc96124a3d66a2a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:33:00 crc kubenswrapper[4706]: E1208 19:33:00.638854 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(6fc26423a36b7251af2cd16592aedcd9217b851e8f334dcecc96124a3d66a2a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:33:00 crc kubenswrapper[4706]: E1208 19:33:00.638877 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(6fc26423a36b7251af2cd16592aedcd9217b851e8f334dcecc96124a3d66a2a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:33:00 crc kubenswrapper[4706]: E1208 19:33:00.638933 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators(f543c2b7-0b00-458c-b4e5-a8c89cfaff5c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators(f543c2b7-0b00-458c-b4e5-a8c89cfaff5c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_openshift-operators_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c_0(6fc26423a36b7251af2cd16592aedcd9217b851e8f334dcecc96124a3d66a2a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" podUID="f543c2b7-0b00-458c-b4e5-a8c89cfaff5c" Dec 08 19:33:00 crc kubenswrapper[4706]: E1208 19:33:00.653762 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(b9b42a56df6933f7804be9799242711d558dd5b67d095f9444e9418541f5b0c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 19:33:00 crc kubenswrapper[4706]: E1208 19:33:00.653855 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(b9b42a56df6933f7804be9799242711d558dd5b67d095f9444e9418541f5b0c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:33:00 crc kubenswrapper[4706]: E1208 19:33:00.653904 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(b9b42a56df6933f7804be9799242711d558dd5b67d095f9444e9418541f5b0c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:33:00 crc kubenswrapper[4706]: E1208 19:33:00.653959 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-fqdzc_openshift-operators(a8e2aac6-1870-4ddc-8650-c31ded55337f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-fqdzc_openshift-operators(a8e2aac6-1870-4ddc-8650-c31ded55337f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fqdzc_openshift-operators_a8e2aac6-1870-4ddc-8650-c31ded55337f_0(b9b42a56df6933f7804be9799242711d558dd5b67d095f9444e9418541f5b0c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" podUID="a8e2aac6-1870-4ddc-8650-c31ded55337f" Dec 08 19:33:03 crc kubenswrapper[4706]: I1208 19:33:03.608901 4706 scope.go:117] "RemoveContainer" containerID="17466bfe4afca9ffa7ad363b6a0c5b80401ec79daedb1e59ad77a6f7bc754587" Dec 08 19:33:04 crc kubenswrapper[4706]: I1208 19:33:04.763784 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x674x_31d41aa9-3a74-42f4-9933-3f484a89c41d/kube-multus/2.log" Dec 08 19:33:04 crc kubenswrapper[4706]: I1208 19:33:04.764197 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x674x" event={"ID":"31d41aa9-3a74-42f4-9933-3f484a89c41d","Type":"ContainerStarted","Data":"4bd8ae3e6ab9709a4d594d2cbdeed226fc47e4377ab27e18cbcc2eea5a95e804"} Dec 08 19:33:09 crc kubenswrapper[4706]: I1208 19:33:09.593983 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r4gr8" Dec 08 19:33:10 crc kubenswrapper[4706]: I1208 19:33:10.607942 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:33:10 crc kubenswrapper[4706]: I1208 19:33:10.608906 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" Dec 08 19:33:10 crc kubenswrapper[4706]: I1208 19:33:10.881914 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx"] Dec 08 19:33:11 crc kubenswrapper[4706]: I1208 19:33:11.607576 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:33:11 crc kubenswrapper[4706]: I1208 19:33:11.608214 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" Dec 08 19:33:11 crc kubenswrapper[4706]: I1208 19:33:11.806583 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" event={"ID":"3f251b1b-6576-4421-aa76-04a26d0e1302","Type":"ContainerStarted","Data":"ce8914e4b07d971ad93daf418b7b612db9b956c0f426233d7b2ea52a2e721260"} Dec 08 19:33:11 crc kubenswrapper[4706]: I1208 19:33:11.841400 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp"] Dec 08 19:33:12 crc kubenswrapper[4706]: I1208 19:33:12.814413 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" event={"ID":"f543c2b7-0b00-458c-b4e5-a8c89cfaff5c","Type":"ContainerStarted","Data":"19dd1159189b4298721bb964e16ec4d628e28f7a9262bfc0dcd8ce9c19db73bd"} Dec 08 19:33:13 crc kubenswrapper[4706]: I1208 19:33:13.608396 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:33:13 crc kubenswrapper[4706]: I1208 19:33:13.608407 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:33:13 crc kubenswrapper[4706]: I1208 19:33:13.609306 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:33:13 crc kubenswrapper[4706]: I1208 19:33:13.609423 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" Dec 08 19:33:13 crc kubenswrapper[4706]: I1208 19:33:13.872323 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr"] Dec 08 19:33:13 crc kubenswrapper[4706]: W1208 19:33:13.891219 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90a15b07_3ae8_4265_a92b_d9dd052211e2.slice/crio-6ffa71a7d459a7c34d93541b3964e57d307280a31a70604af9fc4711012ab75b WatchSource:0}: Error finding container 6ffa71a7d459a7c34d93541b3964e57d307280a31a70604af9fc4711012ab75b: Status 404 returned error can't find the container with id 6ffa71a7d459a7c34d93541b3964e57d307280a31a70604af9fc4711012ab75b Dec 08 19:33:13 crc kubenswrapper[4706]: I1208 19:33:13.905656 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-4xlqn"] Dec 08 19:33:13 crc kubenswrapper[4706]: W1208 19:33:13.922133 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cbb9b0b_a2c7_47f8_92d2_de4829687a45.slice/crio-60a3225ca8291f56b511058fe8e82e9b030e938b1724813a564223c64a72e075 WatchSource:0}: Error finding container 60a3225ca8291f56b511058fe8e82e9b030e938b1724813a564223c64a72e075: Status 404 returned error can't find the container with id 60a3225ca8291f56b511058fe8e82e9b030e938b1724813a564223c64a72e075 Dec 08 19:33:14 crc kubenswrapper[4706]: I1208 19:33:14.608344 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:33:14 crc kubenswrapper[4706]: I1208 19:33:14.609033 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:33:14 crc kubenswrapper[4706]: I1208 19:33:14.841326 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" event={"ID":"90a15b07-3ae8-4265-a92b-d9dd052211e2","Type":"ContainerStarted","Data":"6ffa71a7d459a7c34d93541b3964e57d307280a31a70604af9fc4711012ab75b"} Dec 08 19:33:14 crc kubenswrapper[4706]: I1208 19:33:14.842667 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" event={"ID":"3cbb9b0b-a2c7-47f8-92d2-de4829687a45","Type":"ContainerStarted","Data":"60a3225ca8291f56b511058fe8e82e9b030e938b1724813a564223c64a72e075"} Dec 08 19:33:15 crc kubenswrapper[4706]: I1208 19:33:15.066907 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-fqdzc"] Dec 08 19:33:18 crc kubenswrapper[4706]: I1208 19:33:18.889220 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" event={"ID":"a8e2aac6-1870-4ddc-8650-c31ded55337f","Type":"ContainerStarted","Data":"a6fbc0569238f79d88068875d0bc5a407996b1af4748cbb8724847b002ec4cf4"} Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.901085 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" event={"ID":"a8e2aac6-1870-4ddc-8650-c31ded55337f","Type":"ContainerStarted","Data":"8955d6e9110e5c2523ca1ad507f5f38d6ea785f2db6aa96fd09ca826964f82dc"} Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.901486 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.902555 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" event={"ID":"90a15b07-3ae8-4265-a92b-d9dd052211e2","Type":"ContainerStarted","Data":"6a36645af742133f9ac2aaad4eaa261c4fe07ffadc17b6c739f05d2a4ca3870e"} Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.904319 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" event={"ID":"f543c2b7-0b00-458c-b4e5-a8c89cfaff5c","Type":"ContainerStarted","Data":"825c9d45e9f61a9a4435bea3d416b21d705abb90948fcab47d567704ca8e129c"} Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.905827 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" event={"ID":"3f251b1b-6576-4421-aa76-04a26d0e1302","Type":"ContainerStarted","Data":"f32130133c33a7bf770069f3fc66e123a6f3a158b6fe09740da41d822aeee269"} Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.922607 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" podStartSLOduration=32.242894044 podStartE2EDuration="34.922581786s" podCreationTimestamp="2025-12-08 19:32:46 +0000 UTC" firstStartedPulling="2025-12-08 19:33:18.038471622 +0000 UTC m=+680.680672625" lastFinishedPulling="2025-12-08 19:33:20.718159364 +0000 UTC m=+683.360360367" observedRunningTime="2025-12-08 19:33:20.922572506 +0000 UTC m=+683.564773509" watchObservedRunningTime="2025-12-08 19:33:20.922581786 +0000 UTC m=+683.564782789" Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.965342 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fntdr" podStartSLOduration=29.201619973 podStartE2EDuration="34.965312723s" podCreationTimestamp="2025-12-08 19:32:46 +0000 UTC" firstStartedPulling="2025-12-08 19:33:13.895174668 +0000 UTC m=+676.537375671" lastFinishedPulling="2025-12-08 19:33:19.658867418 +0000 UTC m=+682.301068421" observedRunningTime="2025-12-08 19:33:20.942628657 +0000 UTC m=+683.584829670" watchObservedRunningTime="2025-12-08 19:33:20.965312723 +0000 UTC m=+683.607513726" Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.968852 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-hdhxp" podStartSLOduration=27.131700586 podStartE2EDuration="34.968840013s" podCreationTimestamp="2025-12-08 19:32:46 +0000 UTC" firstStartedPulling="2025-12-08 19:33:11.857642423 +0000 UTC m=+674.499843426" lastFinishedPulling="2025-12-08 19:33:19.69478185 +0000 UTC m=+682.336982853" observedRunningTime="2025-12-08 19:33:20.956824071 +0000 UTC m=+683.599025074" watchObservedRunningTime="2025-12-08 19:33:20.968840013 +0000 UTC m=+683.611041016" Dec 08 19:33:20 crc kubenswrapper[4706]: I1208 19:33:20.995505 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-64d6644894-ctjxx" podStartSLOduration=26.237691036 podStartE2EDuration="34.995482642s" podCreationTimestamp="2025-12-08 19:32:46 +0000 UTC" firstStartedPulling="2025-12-08 19:33:10.891702485 +0000 UTC m=+673.533903488" lastFinishedPulling="2025-12-08 19:33:19.649494091 +0000 UTC m=+682.291695094" observedRunningTime="2025-12-08 19:33:20.992784655 +0000 UTC m=+683.634985658" watchObservedRunningTime="2025-12-08 19:33:20.995482642 +0000 UTC m=+683.637683645" Dec 08 19:33:23 crc kubenswrapper[4706]: I1208 19:33:23.965544 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" event={"ID":"3cbb9b0b-a2c7-47f8-92d2-de4829687a45","Type":"ContainerStarted","Data":"3ad64f4d62b27be591f570cc17c7e973153175ab023199780218d35fcb21682f"} Dec 08 19:33:23 crc kubenswrapper[4706]: I1208 19:33:23.966204 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:33:23 crc kubenswrapper[4706]: I1208 19:33:23.967357 4706 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-4xlqn container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.27:8081/healthz\": dial tcp 10.217.0.27:8081: connect: connection refused" start-of-body= Dec 08 19:33:23 crc kubenswrapper[4706]: I1208 19:33:23.967406 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" podUID="3cbb9b0b-a2c7-47f8-92d2-de4829687a45" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.27:8081/healthz\": dial tcp 10.217.0.27:8081: connect: connection refused" Dec 08 19:33:24 crc kubenswrapper[4706]: I1208 19:33:24.018480 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" podStartSLOduration=28.373283885 podStartE2EDuration="38.018448871s" podCreationTimestamp="2025-12-08 19:32:46 +0000 UTC" firstStartedPulling="2025-12-08 19:33:13.925822831 +0000 UTC m=+676.568023834" lastFinishedPulling="2025-12-08 19:33:23.570987817 +0000 UTC m=+686.213188820" observedRunningTime="2025-12-08 19:33:23.995827176 +0000 UTC m=+686.638028179" watchObservedRunningTime="2025-12-08 19:33:24.018448871 +0000 UTC m=+686.660649884" Dec 08 19:33:25 crc kubenswrapper[4706]: I1208 19:33:25.002064 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-4xlqn" Dec 08 19:33:27 crc kubenswrapper[4706]: I1208 19:33:27.158297 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-fqdzc" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.588903 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hzrrx"] Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.590404 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-hzrrx" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.593376 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.593522 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-j78ck" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.597919 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.606166 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hzrrx"] Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.622064 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-588tt"] Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.623075 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.625060 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-gvm6m" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.633835 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c6s5g"] Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.637484 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-c6s5g" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.641561 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-btmk8" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.650873 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c6s5g"] Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.663502 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-588tt"] Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.688998 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lh25\" (UniqueName: \"kubernetes.io/projected/8bf48929-3e5e-4fc7-a559-0ddcc2707cda-kube-api-access-4lh25\") pod \"cert-manager-5b446d88c5-c6s5g\" (UID: \"8bf48929-3e5e-4fc7-a559-0ddcc2707cda\") " pod="cert-manager/cert-manager-5b446d88c5-c6s5g" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.689096 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d425v\" (UniqueName: \"kubernetes.io/projected/aefedb16-e712-45a1-8c7b-cd34002fd0bb-kube-api-access-d425v\") pod \"cert-manager-cainjector-7f985d654d-hzrrx\" (UID: \"aefedb16-e712-45a1-8c7b-cd34002fd0bb\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hzrrx" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.689123 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kh7p\" (UniqueName: \"kubernetes.io/projected/c1b92b7d-f282-496b-b69f-f55c0b9baec2-kube-api-access-9kh7p\") pod \"cert-manager-webhook-5655c58dd6-588tt\" (UID: \"c1b92b7d-f282-496b-b69f-f55c0b9baec2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.790492 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lh25\" (UniqueName: \"kubernetes.io/projected/8bf48929-3e5e-4fc7-a559-0ddcc2707cda-kube-api-access-4lh25\") pod \"cert-manager-5b446d88c5-c6s5g\" (UID: \"8bf48929-3e5e-4fc7-a559-0ddcc2707cda\") " pod="cert-manager/cert-manager-5b446d88c5-c6s5g" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.790568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d425v\" (UniqueName: \"kubernetes.io/projected/aefedb16-e712-45a1-8c7b-cd34002fd0bb-kube-api-access-d425v\") pod \"cert-manager-cainjector-7f985d654d-hzrrx\" (UID: \"aefedb16-e712-45a1-8c7b-cd34002fd0bb\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hzrrx" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.790599 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kh7p\" (UniqueName: \"kubernetes.io/projected/c1b92b7d-f282-496b-b69f-f55c0b9baec2-kube-api-access-9kh7p\") pod \"cert-manager-webhook-5655c58dd6-588tt\" (UID: \"c1b92b7d-f282-496b-b69f-f55c0b9baec2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.824444 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kh7p\" (UniqueName: \"kubernetes.io/projected/c1b92b7d-f282-496b-b69f-f55c0b9baec2-kube-api-access-9kh7p\") pod \"cert-manager-webhook-5655c58dd6-588tt\" (UID: \"c1b92b7d-f282-496b-b69f-f55c0b9baec2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.825093 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lh25\" (UniqueName: \"kubernetes.io/projected/8bf48929-3e5e-4fc7-a559-0ddcc2707cda-kube-api-access-4lh25\") pod \"cert-manager-5b446d88c5-c6s5g\" (UID: \"8bf48929-3e5e-4fc7-a559-0ddcc2707cda\") " pod="cert-manager/cert-manager-5b446d88c5-c6s5g" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.840249 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d425v\" (UniqueName: \"kubernetes.io/projected/aefedb16-e712-45a1-8c7b-cd34002fd0bb-kube-api-access-d425v\") pod \"cert-manager-cainjector-7f985d654d-hzrrx\" (UID: \"aefedb16-e712-45a1-8c7b-cd34002fd0bb\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hzrrx" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.909653 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-hzrrx" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.937741 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" Dec 08 19:33:31 crc kubenswrapper[4706]: I1208 19:33:31.950660 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-c6s5g" Dec 08 19:33:32 crc kubenswrapper[4706]: I1208 19:33:32.393452 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hzrrx"] Dec 08 19:33:32 crc kubenswrapper[4706]: I1208 19:33:32.712335 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-588tt"] Dec 08 19:33:32 crc kubenswrapper[4706]: I1208 19:33:32.724573 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c6s5g"] Dec 08 19:33:33 crc kubenswrapper[4706]: I1208 19:33:33.024878 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" event={"ID":"c1b92b7d-f282-496b-b69f-f55c0b9baec2","Type":"ContainerStarted","Data":"4086cd136aa517bca4ac68e489363164b327753775443ba0838e63e0096fd6fb"} Dec 08 19:33:33 crc kubenswrapper[4706]: I1208 19:33:33.026038 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-hzrrx" event={"ID":"aefedb16-e712-45a1-8c7b-cd34002fd0bb","Type":"ContainerStarted","Data":"8f8a0b0f9a3e73908ae9d1c5ca504cb1d4b6b690066dc991329c386c641fde82"} Dec 08 19:33:33 crc kubenswrapper[4706]: I1208 19:33:33.027795 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-c6s5g" event={"ID":"8bf48929-3e5e-4fc7-a559-0ddcc2707cda","Type":"ContainerStarted","Data":"78e3e266e926f682f25e7ebfbaa335d92ce14db8eed33a70596e38a2f59af441"} Dec 08 19:33:36 crc kubenswrapper[4706]: I1208 19:33:36.047721 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" event={"ID":"c1b92b7d-f282-496b-b69f-f55c0b9baec2","Type":"ContainerStarted","Data":"d0bf5920d924e729cf36703afa2148a1160df3ff083db3b0c081479765d5a8d1"} Dec 08 19:33:36 crc kubenswrapper[4706]: I1208 19:33:36.048150 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" Dec 08 19:33:36 crc kubenswrapper[4706]: I1208 19:33:36.049565 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-hzrrx" event={"ID":"aefedb16-e712-45a1-8c7b-cd34002fd0bb","Type":"ContainerStarted","Data":"63cfdadbcc0dc9d60388e27ec486bed4f432cc69cc954f3f8c7fecab5f115d2d"} Dec 08 19:33:36 crc kubenswrapper[4706]: I1208 19:33:36.104044 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-hzrrx" podStartSLOduration=2.346789555 podStartE2EDuration="5.104025457s" podCreationTimestamp="2025-12-08 19:33:31 +0000 UTC" firstStartedPulling="2025-12-08 19:33:32.408765622 +0000 UTC m=+695.050966625" lastFinishedPulling="2025-12-08 19:33:35.166001524 +0000 UTC m=+697.808202527" observedRunningTime="2025-12-08 19:33:36.102596096 +0000 UTC m=+698.744797099" watchObservedRunningTime="2025-12-08 19:33:36.104025457 +0000 UTC m=+698.746226460" Dec 08 19:33:36 crc kubenswrapper[4706]: I1208 19:33:36.107580 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" podStartSLOduration=2.597138486 podStartE2EDuration="5.107571168s" podCreationTimestamp="2025-12-08 19:33:31 +0000 UTC" firstStartedPulling="2025-12-08 19:33:32.72013366 +0000 UTC m=+695.362334663" lastFinishedPulling="2025-12-08 19:33:35.230566332 +0000 UTC m=+697.872767345" observedRunningTime="2025-12-08 19:33:36.084484841 +0000 UTC m=+698.726685844" watchObservedRunningTime="2025-12-08 19:33:36.107571168 +0000 UTC m=+698.749772171" Dec 08 19:33:37 crc kubenswrapper[4706]: I1208 19:33:37.057568 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-c6s5g" event={"ID":"8bf48929-3e5e-4fc7-a559-0ddcc2707cda","Type":"ContainerStarted","Data":"f39677239fe468bfaa75984318cb5e90497b30ed70835151ab2e35e23dabc0c0"} Dec 08 19:33:37 crc kubenswrapper[4706]: I1208 19:33:37.100453 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-c6s5g" podStartSLOduration=2.259309374 podStartE2EDuration="6.100422483s" podCreationTimestamp="2025-12-08 19:33:31 +0000 UTC" firstStartedPulling="2025-12-08 19:33:32.736448904 +0000 UTC m=+695.378649907" lastFinishedPulling="2025-12-08 19:33:36.577562013 +0000 UTC m=+699.219763016" observedRunningTime="2025-12-08 19:33:37.086420984 +0000 UTC m=+699.728621997" watchObservedRunningTime="2025-12-08 19:33:37.100422483 +0000 UTC m=+699.742623486" Dec 08 19:33:41 crc kubenswrapper[4706]: I1208 19:33:41.943288 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-588tt" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.230116 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd"] Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.233635 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.238732 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.240071 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd"] Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.323167 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjpdn\" (UniqueName: \"kubernetes.io/projected/b0c2a5d5-da86-47ef-9785-323261350321-kube-api-access-tjpdn\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.323220 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-bundle\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.323274 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-util\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.424223 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjpdn\" (UniqueName: \"kubernetes.io/projected/b0c2a5d5-da86-47ef-9785-323261350321-kube-api-access-tjpdn\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.424293 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-bundle\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.424319 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-util\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.424849 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-util\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.425392 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-bundle\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.445771 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjpdn\" (UniqueName: \"kubernetes.io/projected/b0c2a5d5-da86-47ef-9785-323261350321-kube-api-access-tjpdn\") pod \"7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.560083 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:07 crc kubenswrapper[4706]: I1208 19:34:07.808463 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd"] Dec 08 19:34:08 crc kubenswrapper[4706]: I1208 19:34:08.262284 4706 generic.go:334] "Generic (PLEG): container finished" podID="b0c2a5d5-da86-47ef-9785-323261350321" containerID="26485a548320e448d59d510d52ce484d85b3a1f387be18aa16bd26bff0c819a8" exitCode=0 Dec 08 19:34:08 crc kubenswrapper[4706]: I1208 19:34:08.262343 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" event={"ID":"b0c2a5d5-da86-47ef-9785-323261350321","Type":"ContainerDied","Data":"26485a548320e448d59d510d52ce484d85b3a1f387be18aa16bd26bff0c819a8"} Dec 08 19:34:08 crc kubenswrapper[4706]: I1208 19:34:08.262377 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" event={"ID":"b0c2a5d5-da86-47ef-9785-323261350321","Type":"ContainerStarted","Data":"2311234deb38e6de3ae5ead65b10ce2231877146648dca7d67e5ae7359e1df7d"} Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.240412 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.241625 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.243805 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.244345 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.254487 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.294820 4706 generic.go:334] "Generic (PLEG): container finished" podID="b0c2a5d5-da86-47ef-9785-323261350321" containerID="31b2a919f9cfc1436fb4f3124e627098c1696d63298071f1547f3de8b8422b36" exitCode=0 Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.294869 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" event={"ID":"b0c2a5d5-da86-47ef-9785-323261350321","Type":"ContainerDied","Data":"31b2a919f9cfc1436fb4f3124e627098c1696d63298071f1547f3de8b8422b36"} Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.361509 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b7a0775b-1fae-4fda-85fd-a41d7b549aaa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b7a0775b-1fae-4fda-85fd-a41d7b549aaa\") pod \"minio\" (UID: \"32f51328-fa11-4d43-9ac3-bc93e92c3149\") " pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.361944 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72bz2\" (UniqueName: \"kubernetes.io/projected/32f51328-fa11-4d43-9ac3-bc93e92c3149-kube-api-access-72bz2\") pod \"minio\" (UID: \"32f51328-fa11-4d43-9ac3-bc93e92c3149\") " pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.463665 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72bz2\" (UniqueName: \"kubernetes.io/projected/32f51328-fa11-4d43-9ac3-bc93e92c3149-kube-api-access-72bz2\") pod \"minio\" (UID: \"32f51328-fa11-4d43-9ac3-bc93e92c3149\") " pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.463805 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b7a0775b-1fae-4fda-85fd-a41d7b549aaa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b7a0775b-1fae-4fda-85fd-a41d7b549aaa\") pod \"minio\" (UID: \"32f51328-fa11-4d43-9ac3-bc93e92c3149\") " pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.472179 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.472250 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b7a0775b-1fae-4fda-85fd-a41d7b549aaa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b7a0775b-1fae-4fda-85fd-a41d7b549aaa\") pod \"minio\" (UID: \"32f51328-fa11-4d43-9ac3-bc93e92c3149\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6c7a07fb7d2e04c6bd872586e677d1c627383ffb855f0c6021ed2f10d8b7b494/globalmount\"" pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.490975 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72bz2\" (UniqueName: \"kubernetes.io/projected/32f51328-fa11-4d43-9ac3-bc93e92c3149-kube-api-access-72bz2\") pod \"minio\" (UID: \"32f51328-fa11-4d43-9ac3-bc93e92c3149\") " pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.512503 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b7a0775b-1fae-4fda-85fd-a41d7b549aaa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b7a0775b-1fae-4fda-85fd-a41d7b549aaa\") pod \"minio\" (UID: \"32f51328-fa11-4d43-9ac3-bc93e92c3149\") " pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.559342 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 08 19:34:10 crc kubenswrapper[4706]: I1208 19:34:10.763673 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 08 19:34:10 crc kubenswrapper[4706]: W1208 19:34:10.780680 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32f51328_fa11_4d43_9ac3_bc93e92c3149.slice/crio-00d74ee4ae6d7aff2bbaf76a27d679f16b619008f5cbc91e238085d4fb519544 WatchSource:0}: Error finding container 00d74ee4ae6d7aff2bbaf76a27d679f16b619008f5cbc91e238085d4fb519544: Status 404 returned error can't find the container with id 00d74ee4ae6d7aff2bbaf76a27d679f16b619008f5cbc91e238085d4fb519544 Dec 08 19:34:11 crc kubenswrapper[4706]: I1208 19:34:11.303774 4706 generic.go:334] "Generic (PLEG): container finished" podID="b0c2a5d5-da86-47ef-9785-323261350321" containerID="7c185ae53f5c8b12f341cf3efbb4f57e4716386f5cfa62ce1342ddacfe7325c5" exitCode=0 Dec 08 19:34:11 crc kubenswrapper[4706]: I1208 19:34:11.303860 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" event={"ID":"b0c2a5d5-da86-47ef-9785-323261350321","Type":"ContainerDied","Data":"7c185ae53f5c8b12f341cf3efbb4f57e4716386f5cfa62ce1342ddacfe7325c5"} Dec 08 19:34:11 crc kubenswrapper[4706]: I1208 19:34:11.304959 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"32f51328-fa11-4d43-9ac3-bc93e92c3149","Type":"ContainerStarted","Data":"00d74ee4ae6d7aff2bbaf76a27d679f16b619008f5cbc91e238085d4fb519544"} Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.513080 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.615628 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjpdn\" (UniqueName: \"kubernetes.io/projected/b0c2a5d5-da86-47ef-9785-323261350321-kube-api-access-tjpdn\") pod \"b0c2a5d5-da86-47ef-9785-323261350321\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.616092 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-util\") pod \"b0c2a5d5-da86-47ef-9785-323261350321\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.616120 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-bundle\") pod \"b0c2a5d5-da86-47ef-9785-323261350321\" (UID: \"b0c2a5d5-da86-47ef-9785-323261350321\") " Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.619362 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-bundle" (OuterVolumeSpecName: "bundle") pod "b0c2a5d5-da86-47ef-9785-323261350321" (UID: "b0c2a5d5-da86-47ef-9785-323261350321"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.624780 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0c2a5d5-da86-47ef-9785-323261350321-kube-api-access-tjpdn" (OuterVolumeSpecName: "kube-api-access-tjpdn") pod "b0c2a5d5-da86-47ef-9785-323261350321" (UID: "b0c2a5d5-da86-47ef-9785-323261350321"). InnerVolumeSpecName "kube-api-access-tjpdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.720147 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-util" (OuterVolumeSpecName: "util") pod "b0c2a5d5-da86-47ef-9785-323261350321" (UID: "b0c2a5d5-da86-47ef-9785-323261350321"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.721151 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-util\") on node \"crc\" DevicePath \"\"" Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.721178 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0c2a5d5-da86-47ef-9785-323261350321-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:34:13 crc kubenswrapper[4706]: I1208 19:34:13.721191 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjpdn\" (UniqueName: \"kubernetes.io/projected/b0c2a5d5-da86-47ef-9785-323261350321-kube-api-access-tjpdn\") on node \"crc\" DevicePath \"\"" Dec 08 19:34:14 crc kubenswrapper[4706]: I1208 19:34:14.327233 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" event={"ID":"b0c2a5d5-da86-47ef-9785-323261350321","Type":"ContainerDied","Data":"2311234deb38e6de3ae5ead65b10ce2231877146648dca7d67e5ae7359e1df7d"} Dec 08 19:34:14 crc kubenswrapper[4706]: I1208 19:34:14.327306 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2311234deb38e6de3ae5ead65b10ce2231877146648dca7d67e5ae7359e1df7d" Dec 08 19:34:14 crc kubenswrapper[4706]: I1208 19:34:14.327406 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd" Dec 08 19:34:15 crc kubenswrapper[4706]: I1208 19:34:15.334940 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"32f51328-fa11-4d43-9ac3-bc93e92c3149","Type":"ContainerStarted","Data":"85a61fc72f0720a384e6bc180b3cf41e087c2959979129b85f0b67d53259cda6"} Dec 08 19:34:15 crc kubenswrapper[4706]: I1208 19:34:15.359946 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.818565173 podStartE2EDuration="8.359923559s" podCreationTimestamp="2025-12-08 19:34:07 +0000 UTC" firstStartedPulling="2025-12-08 19:34:10.784380259 +0000 UTC m=+733.426581262" lastFinishedPulling="2025-12-08 19:34:14.325738645 +0000 UTC m=+736.967939648" observedRunningTime="2025-12-08 19:34:15.356177813 +0000 UTC m=+737.998378826" watchObservedRunningTime="2025-12-08 19:34:15.359923559 +0000 UTC m=+738.002124562" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.170568 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q"] Dec 08 19:34:19 crc kubenswrapper[4706]: E1208 19:34:19.171202 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c2a5d5-da86-47ef-9785-323261350321" containerName="pull" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.171217 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c2a5d5-da86-47ef-9785-323261350321" containerName="pull" Dec 08 19:34:19 crc kubenswrapper[4706]: E1208 19:34:19.171228 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c2a5d5-da86-47ef-9785-323261350321" containerName="util" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.171235 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c2a5d5-da86-47ef-9785-323261350321" containerName="util" Dec 08 19:34:19 crc kubenswrapper[4706]: E1208 19:34:19.171271 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c2a5d5-da86-47ef-9785-323261350321" containerName="extract" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.171279 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c2a5d5-da86-47ef-9785-323261350321" containerName="extract" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.171387 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0c2a5d5-da86-47ef-9785-323261350321" containerName="extract" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.172035 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.176596 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.176721 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.176798 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.176858 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-245v5" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.181195 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.181241 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.194627 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q"] Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.305140 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b600c67d-b283-41ed-85f1-d19f3671f40f-manager-config\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.305557 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr86l\" (UniqueName: \"kubernetes.io/projected/b600c67d-b283-41ed-85f1-d19f3671f40f-kube-api-access-zr86l\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.305868 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-apiservice-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.305928 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-webhook-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.305998 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.407194 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b600c67d-b283-41ed-85f1-d19f3671f40f-manager-config\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.407293 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr86l\" (UniqueName: \"kubernetes.io/projected/b600c67d-b283-41ed-85f1-d19f3671f40f-kube-api-access-zr86l\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.407336 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-apiservice-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.407356 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-webhook-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.407385 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.408590 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b600c67d-b283-41ed-85f1-d19f3671f40f-manager-config\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.414254 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-webhook-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.418238 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-apiservice-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.424296 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b600c67d-b283-41ed-85f1-d19f3671f40f-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.440639 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr86l\" (UniqueName: \"kubernetes.io/projected/b600c67d-b283-41ed-85f1-d19f3671f40f-kube-api-access-zr86l\") pod \"loki-operator-controller-manager-55474f77b6-v6n6q\" (UID: \"b600c67d-b283-41ed-85f1-d19f3671f40f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.490883 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:19 crc kubenswrapper[4706]: I1208 19:34:19.740300 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q"] Dec 08 19:34:19 crc kubenswrapper[4706]: W1208 19:34:19.746919 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb600c67d_b283_41ed_85f1_d19f3671f40f.slice/crio-252a4045d4306ce31c97afea872ca27e981c3f42f4d47037296d746c030937f0 WatchSource:0}: Error finding container 252a4045d4306ce31c97afea872ca27e981c3f42f4d47037296d746c030937f0: Status 404 returned error can't find the container with id 252a4045d4306ce31c97afea872ca27e981c3f42f4d47037296d746c030937f0 Dec 08 19:34:20 crc kubenswrapper[4706]: I1208 19:34:20.366184 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" event={"ID":"b600c67d-b283-41ed-85f1-d19f3671f40f","Type":"ContainerStarted","Data":"252a4045d4306ce31c97afea872ca27e981c3f42f4d47037296d746c030937f0"} Dec 08 19:34:25 crc kubenswrapper[4706]: I1208 19:34:25.406467 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" event={"ID":"b600c67d-b283-41ed-85f1-d19f3671f40f","Type":"ContainerStarted","Data":"7cf2eb8704747769a5c9c4df0f4f2040eb998ce1ebc8bd726ca0679f8249864a"} Dec 08 19:34:32 crc kubenswrapper[4706]: I1208 19:34:32.459585 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" event={"ID":"b600c67d-b283-41ed-85f1-d19f3671f40f","Type":"ContainerStarted","Data":"47d3d07677c776cbe14727286f384c1e47d203dd397d33972120d8d957240030"} Dec 08 19:34:32 crc kubenswrapper[4706]: I1208 19:34:32.460527 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:32 crc kubenswrapper[4706]: I1208 19:34:32.463252 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" Dec 08 19:34:32 crc kubenswrapper[4706]: I1208 19:34:32.489188 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-55474f77b6-v6n6q" podStartSLOduration=0.942799777 podStartE2EDuration="13.489168471s" podCreationTimestamp="2025-12-08 19:34:19 +0000 UTC" firstStartedPulling="2025-12-08 19:34:19.749515529 +0000 UTC m=+742.391716532" lastFinishedPulling="2025-12-08 19:34:32.295884233 +0000 UTC m=+754.938085226" observedRunningTime="2025-12-08 19:34:32.486862955 +0000 UTC m=+755.129063968" watchObservedRunningTime="2025-12-08 19:34:32.489168471 +0000 UTC m=+755.131369474" Dec 08 19:34:39 crc kubenswrapper[4706]: I1208 19:34:39.129029 4706 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 08 19:35:05 crc kubenswrapper[4706]: I1208 19:35:05.836614 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:35:05 crc kubenswrapper[4706]: I1208 19:35:05.837327 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.006356 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv"] Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.007649 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.010197 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.018213 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv"] Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.073655 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.073748 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.073801 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndcjg\" (UniqueName: \"kubernetes.io/projected/03a7459f-3dd0-4cc8-9183-b82974d15a55-kube-api-access-ndcjg\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.175079 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndcjg\" (UniqueName: \"kubernetes.io/projected/03a7459f-3dd0-4cc8-9183-b82974d15a55-kube-api-access-ndcjg\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.175200 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.175251 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.175944 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.175956 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.201608 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndcjg\" (UniqueName: \"kubernetes.io/projected/03a7459f-3dd0-4cc8-9183-b82974d15a55-kube-api-access-ndcjg\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.327481 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:06 crc kubenswrapper[4706]: I1208 19:35:06.806039 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv"] Dec 08 19:35:07 crc kubenswrapper[4706]: I1208 19:35:07.669067 4706 generic.go:334] "Generic (PLEG): container finished" podID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerID="f3620546e4c6bc533efa60e57d2d8aa4718b3080f1fefe9eff3609dfbede1395" exitCode=0 Dec 08 19:35:07 crc kubenswrapper[4706]: I1208 19:35:07.669137 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" event={"ID":"03a7459f-3dd0-4cc8-9183-b82974d15a55","Type":"ContainerDied","Data":"f3620546e4c6bc533efa60e57d2d8aa4718b3080f1fefe9eff3609dfbede1395"} Dec 08 19:35:07 crc kubenswrapper[4706]: I1208 19:35:07.669513 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" event={"ID":"03a7459f-3dd0-4cc8-9183-b82974d15a55","Type":"ContainerStarted","Data":"ebffac9e678aa57627c665abe3c676c0bde1038baa6408185b89c6785e677e43"} Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.378088 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kxqlk"] Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.379710 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.397068 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxqlk"] Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.411004 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hxnb\" (UniqueName: \"kubernetes.io/projected/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-kube-api-access-5hxnb\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.411061 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-utilities\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.411094 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-catalog-content\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.512965 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hxnb\" (UniqueName: \"kubernetes.io/projected/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-kube-api-access-5hxnb\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.513425 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-catalog-content\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.513451 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-utilities\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.513968 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-catalog-content\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.514044 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-utilities\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.536610 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hxnb\" (UniqueName: \"kubernetes.io/projected/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-kube-api-access-5hxnb\") pod \"redhat-operators-kxqlk\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:08 crc kubenswrapper[4706]: I1208 19:35:08.715333 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:09 crc kubenswrapper[4706]: I1208 19:35:09.014155 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxqlk"] Dec 08 19:35:09 crc kubenswrapper[4706]: I1208 19:35:09.682839 4706 generic.go:334] "Generic (PLEG): container finished" podID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerID="0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af" exitCode=0 Dec 08 19:35:09 crc kubenswrapper[4706]: I1208 19:35:09.682935 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxqlk" event={"ID":"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6","Type":"ContainerDied","Data":"0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af"} Dec 08 19:35:09 crc kubenswrapper[4706]: I1208 19:35:09.683590 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxqlk" event={"ID":"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6","Type":"ContainerStarted","Data":"b0fdf15898598676b7256f61ba0e2b7472472f1985660fcf41b41760bda0a192"} Dec 08 19:35:09 crc kubenswrapper[4706]: I1208 19:35:09.685989 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" event={"ID":"03a7459f-3dd0-4cc8-9183-b82974d15a55","Type":"ContainerStarted","Data":"5ca97d386a4f157e989a04f874a222a1ad530770a153fd3db6dfad2eebb6e7db"} Dec 08 19:35:10 crc kubenswrapper[4706]: I1208 19:35:10.694899 4706 generic.go:334] "Generic (PLEG): container finished" podID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerID="5ca97d386a4f157e989a04f874a222a1ad530770a153fd3db6dfad2eebb6e7db" exitCode=0 Dec 08 19:35:10 crc kubenswrapper[4706]: I1208 19:35:10.694988 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" event={"ID":"03a7459f-3dd0-4cc8-9183-b82974d15a55","Type":"ContainerDied","Data":"5ca97d386a4f157e989a04f874a222a1ad530770a153fd3db6dfad2eebb6e7db"} Dec 08 19:35:10 crc kubenswrapper[4706]: I1208 19:35:10.700612 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxqlk" event={"ID":"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6","Type":"ContainerStarted","Data":"2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193"} Dec 08 19:35:11 crc kubenswrapper[4706]: I1208 19:35:11.709814 4706 generic.go:334] "Generic (PLEG): container finished" podID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerID="350c969c0c92cff03fea767dbeb222adb886b83bdf1baff6520bd712dba40796" exitCode=0 Dec 08 19:35:11 crc kubenswrapper[4706]: I1208 19:35:11.709898 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" event={"ID":"03a7459f-3dd0-4cc8-9183-b82974d15a55","Type":"ContainerDied","Data":"350c969c0c92cff03fea767dbeb222adb886b83bdf1baff6520bd712dba40796"} Dec 08 19:35:12 crc kubenswrapper[4706]: I1208 19:35:12.716878 4706 generic.go:334] "Generic (PLEG): container finished" podID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerID="2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193" exitCode=0 Dec 08 19:35:12 crc kubenswrapper[4706]: I1208 19:35:12.716945 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxqlk" event={"ID":"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6","Type":"ContainerDied","Data":"2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193"} Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.052189 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.183467 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-bundle\") pod \"03a7459f-3dd0-4cc8-9183-b82974d15a55\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.183522 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-util\") pod \"03a7459f-3dd0-4cc8-9183-b82974d15a55\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.183582 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndcjg\" (UniqueName: \"kubernetes.io/projected/03a7459f-3dd0-4cc8-9183-b82974d15a55-kube-api-access-ndcjg\") pod \"03a7459f-3dd0-4cc8-9183-b82974d15a55\" (UID: \"03a7459f-3dd0-4cc8-9183-b82974d15a55\") " Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.184160 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-bundle" (OuterVolumeSpecName: "bundle") pod "03a7459f-3dd0-4cc8-9183-b82974d15a55" (UID: "03a7459f-3dd0-4cc8-9183-b82974d15a55"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.190611 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03a7459f-3dd0-4cc8-9183-b82974d15a55-kube-api-access-ndcjg" (OuterVolumeSpecName: "kube-api-access-ndcjg") pod "03a7459f-3dd0-4cc8-9183-b82974d15a55" (UID: "03a7459f-3dd0-4cc8-9183-b82974d15a55"). InnerVolumeSpecName "kube-api-access-ndcjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.194713 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-util" (OuterVolumeSpecName: "util") pod "03a7459f-3dd0-4cc8-9183-b82974d15a55" (UID: "03a7459f-3dd0-4cc8-9183-b82974d15a55"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.285729 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.285785 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/03a7459f-3dd0-4cc8-9183-b82974d15a55-util\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.285798 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndcjg\" (UniqueName: \"kubernetes.io/projected/03a7459f-3dd0-4cc8-9183-b82974d15a55-kube-api-access-ndcjg\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.726290 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" event={"ID":"03a7459f-3dd0-4cc8-9183-b82974d15a55","Type":"ContainerDied","Data":"ebffac9e678aa57627c665abe3c676c0bde1038baa6408185b89c6785e677e43"} Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.726324 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.726337 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebffac9e678aa57627c665abe3c676c0bde1038baa6408185b89c6785e677e43" Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.728542 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxqlk" event={"ID":"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6","Type":"ContainerStarted","Data":"4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a"} Dec 08 19:35:13 crc kubenswrapper[4706]: I1208 19:35:13.752049 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kxqlk" podStartSLOduration=2.131298878 podStartE2EDuration="5.752020639s" podCreationTimestamp="2025-12-08 19:35:08 +0000 UTC" firstStartedPulling="2025-12-08 19:35:09.685170398 +0000 UTC m=+792.327371401" lastFinishedPulling="2025-12-08 19:35:13.305892159 +0000 UTC m=+795.948093162" observedRunningTime="2025-12-08 19:35:13.746885494 +0000 UTC m=+796.389086497" watchObservedRunningTime="2025-12-08 19:35:13.752020639 +0000 UTC m=+796.394221652" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.808170 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k"] Dec 08 19:35:15 crc kubenswrapper[4706]: E1208 19:35:15.808856 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerName="pull" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.808872 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerName="pull" Dec 08 19:35:15 crc kubenswrapper[4706]: E1208 19:35:15.808887 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerName="util" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.808893 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerName="util" Dec 08 19:35:15 crc kubenswrapper[4706]: E1208 19:35:15.808904 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerName="extract" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.808914 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerName="extract" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.809024 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="03a7459f-3dd0-4cc8-9183-b82974d15a55" containerName="extract" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.809597 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.812794 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.812815 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.813750 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-vvpkn" Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.831650 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k"] Dec 08 19:35:15 crc kubenswrapper[4706]: I1208 19:35:15.921371 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfl6k\" (UniqueName: \"kubernetes.io/projected/f56b4031-2db2-4d47-9299-9829b3963664-kube-api-access-vfl6k\") pod \"nmstate-operator-5b5b58f5c8-9jz6k\" (UID: \"f56b4031-2db2-4d47-9299-9829b3963664\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k" Dec 08 19:35:16 crc kubenswrapper[4706]: I1208 19:35:16.023085 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfl6k\" (UniqueName: \"kubernetes.io/projected/f56b4031-2db2-4d47-9299-9829b3963664-kube-api-access-vfl6k\") pod \"nmstate-operator-5b5b58f5c8-9jz6k\" (UID: \"f56b4031-2db2-4d47-9299-9829b3963664\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k" Dec 08 19:35:16 crc kubenswrapper[4706]: I1208 19:35:16.043586 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfl6k\" (UniqueName: \"kubernetes.io/projected/f56b4031-2db2-4d47-9299-9829b3963664-kube-api-access-vfl6k\") pod \"nmstate-operator-5b5b58f5c8-9jz6k\" (UID: \"f56b4031-2db2-4d47-9299-9829b3963664\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k" Dec 08 19:35:16 crc kubenswrapper[4706]: I1208 19:35:16.126795 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k" Dec 08 19:35:16 crc kubenswrapper[4706]: I1208 19:35:16.408182 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k"] Dec 08 19:35:16 crc kubenswrapper[4706]: I1208 19:35:16.746980 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k" event={"ID":"f56b4031-2db2-4d47-9299-9829b3963664","Type":"ContainerStarted","Data":"32af1ad4da7e459457007304c87b4501ff422ed130907a9a62183d78c965556c"} Dec 08 19:35:18 crc kubenswrapper[4706]: I1208 19:35:18.716533 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:18 crc kubenswrapper[4706]: I1208 19:35:18.716979 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:19 crc kubenswrapper[4706]: I1208 19:35:19.763360 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kxqlk" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="registry-server" probeResult="failure" output=< Dec 08 19:35:19 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:35:19 crc kubenswrapper[4706]: > Dec 08 19:35:19 crc kubenswrapper[4706]: I1208 19:35:19.770922 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k" event={"ID":"f56b4031-2db2-4d47-9299-9829b3963664","Type":"ContainerStarted","Data":"d587b6936ed4564d974b92d1c2713dfd98dc882e264c276acab0088369da938d"} Dec 08 19:35:19 crc kubenswrapper[4706]: I1208 19:35:19.790596 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9jz6k" podStartSLOduration=1.952248774 podStartE2EDuration="4.790574913s" podCreationTimestamp="2025-12-08 19:35:15 +0000 UTC" firstStartedPulling="2025-12-08 19:35:16.426575634 +0000 UTC m=+799.068776637" lastFinishedPulling="2025-12-08 19:35:19.264901773 +0000 UTC m=+801.907102776" observedRunningTime="2025-12-08 19:35:19.789848472 +0000 UTC m=+802.432049475" watchObservedRunningTime="2025-12-08 19:35:19.790574913 +0000 UTC m=+802.432775916" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.756599 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx"] Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.757739 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.762762 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-sgz74" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.778203 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx"] Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.785633 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx"] Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.787003 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.814149 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.824232 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx"] Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.835329 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-zqns4"] Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.836491 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.894875 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aea9659b-8952-4f71-a8d5-970bfd9fffe0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-kwnzx\" (UID: \"aea9659b-8952-4f71-a8d5-970bfd9fffe0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.895029 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6sqf\" (UniqueName: \"kubernetes.io/projected/aea9659b-8952-4f71-a8d5-970bfd9fffe0-kube-api-access-s6sqf\") pod \"nmstate-webhook-5f6d4c5ccb-kwnzx\" (UID: \"aea9659b-8952-4f71-a8d5-970bfd9fffe0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.895220 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hwhc\" (UniqueName: \"kubernetes.io/projected/05ea3b8c-286e-48e7-b232-ac550fcf165e-kube-api-access-6hwhc\") pod \"nmstate-metrics-7f946cbc9-llfnx\" (UID: \"05ea3b8c-286e-48e7-b232-ac550fcf165e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.968068 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh"] Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.969110 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.974960 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.975092 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.975151 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-l4grm" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.986675 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh"] Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.996215 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aea9659b-8952-4f71-a8d5-970bfd9fffe0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-kwnzx\" (UID: \"aea9659b-8952-4f71-a8d5-970bfd9fffe0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.996317 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-dbus-socket\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.996355 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6sqf\" (UniqueName: \"kubernetes.io/projected/aea9659b-8952-4f71-a8d5-970bfd9fffe0-kube-api-access-s6sqf\") pod \"nmstate-webhook-5f6d4c5ccb-kwnzx\" (UID: \"aea9659b-8952-4f71-a8d5-970bfd9fffe0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.996407 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-nmstate-lock\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:20 crc kubenswrapper[4706]: E1208 19:35:20.996429 4706 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.996437 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-ovs-socket\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:20 crc kubenswrapper[4706]: E1208 19:35:20.996530 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aea9659b-8952-4f71-a8d5-970bfd9fffe0-tls-key-pair podName:aea9659b-8952-4f71-a8d5-970bfd9fffe0 nodeName:}" failed. No retries permitted until 2025-12-08 19:35:21.496502056 +0000 UTC m=+804.138703119 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/aea9659b-8952-4f71-a8d5-970bfd9fffe0-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-kwnzx" (UID: "aea9659b-8952-4f71-a8d5-970bfd9fffe0") : secret "openshift-nmstate-webhook" not found Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.996628 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hwhc\" (UniqueName: \"kubernetes.io/projected/05ea3b8c-286e-48e7-b232-ac550fcf165e-kube-api-access-6hwhc\") pod \"nmstate-metrics-7f946cbc9-llfnx\" (UID: \"05ea3b8c-286e-48e7-b232-ac550fcf165e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" Dec 08 19:35:20 crc kubenswrapper[4706]: I1208 19:35:20.996671 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vsm8\" (UniqueName: \"kubernetes.io/projected/1287503f-533b-4f53-b9b2-6792ad891790-kube-api-access-7vsm8\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.030164 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hwhc\" (UniqueName: \"kubernetes.io/projected/05ea3b8c-286e-48e7-b232-ac550fcf165e-kube-api-access-6hwhc\") pod \"nmstate-metrics-7f946cbc9-llfnx\" (UID: \"05ea3b8c-286e-48e7-b232-ac550fcf165e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.037006 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6sqf\" (UniqueName: \"kubernetes.io/projected/aea9659b-8952-4f71-a8d5-970bfd9fffe0-kube-api-access-s6sqf\") pod \"nmstate-webhook-5f6d4c5ccb-kwnzx\" (UID: \"aea9659b-8952-4f71-a8d5-970bfd9fffe0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.076178 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098435 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-dbus-socket\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098524 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b6da606-8fe3-4228-81f7-533d8155c6f0-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098558 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtfz4\" (UniqueName: \"kubernetes.io/projected/0b6da606-8fe3-4228-81f7-533d8155c6f0-kube-api-access-jtfz4\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098589 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0b6da606-8fe3-4228-81f7-533d8155c6f0-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098612 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-nmstate-lock\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098628 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-ovs-socket\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098718 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vsm8\" (UniqueName: \"kubernetes.io/projected/1287503f-533b-4f53-b9b2-6792ad891790-kube-api-access-7vsm8\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098721 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-nmstate-lock\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.098908 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-dbus-socket\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.099790 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1287503f-533b-4f53-b9b2-6792ad891790-ovs-socket\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.123639 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vsm8\" (UniqueName: \"kubernetes.io/projected/1287503f-533b-4f53-b9b2-6792ad891790-kube-api-access-7vsm8\") pod \"nmstate-handler-zqns4\" (UID: \"1287503f-533b-4f53-b9b2-6792ad891790\") " pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.162311 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.185466 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-75d6767b4d-z8c8m"] Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.186345 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.200240 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b6da606-8fe3-4228-81f7-533d8155c6f0-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.200303 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtfz4\" (UniqueName: \"kubernetes.io/projected/0b6da606-8fe3-4228-81f7-533d8155c6f0-kube-api-access-jtfz4\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.200324 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0b6da606-8fe3-4228-81f7-533d8155c6f0-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.201566 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0b6da606-8fe3-4228-81f7-533d8155c6f0-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.203357 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75d6767b4d-z8c8m"] Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.210173 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b6da606-8fe3-4228-81f7-533d8155c6f0-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.227068 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtfz4\" (UniqueName: \"kubernetes.io/projected/0b6da606-8fe3-4228-81f7-533d8155c6f0-kube-api-access-jtfz4\") pod \"nmstate-console-plugin-7fbb5f6569-224bh\" (UID: \"0b6da606-8fe3-4228-81f7-533d8155c6f0\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.288899 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.302930 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gqvn\" (UniqueName: \"kubernetes.io/projected/abc48bbe-a029-4324-b944-54e83092b7c2-kube-api-access-9gqvn\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.303074 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-trusted-ca-bundle\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.303343 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-oauth-serving-cert\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.304397 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/abc48bbe-a029-4324-b944-54e83092b7c2-console-oauth-config\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.304784 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-service-ca\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.304935 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/abc48bbe-a029-4324-b944-54e83092b7c2-console-serving-cert\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.304991 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-console-config\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.405610 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gqvn\" (UniqueName: \"kubernetes.io/projected/abc48bbe-a029-4324-b944-54e83092b7c2-kube-api-access-9gqvn\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.405669 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-oauth-serving-cert\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.405698 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/abc48bbe-a029-4324-b944-54e83092b7c2-console-oauth-config\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.405715 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-trusted-ca-bundle\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.405736 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-service-ca\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.405777 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/abc48bbe-a029-4324-b944-54e83092b7c2-console-serving-cert\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.405798 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-console-config\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.406679 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-console-config\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.407497 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-oauth-serving-cert\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.408622 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-service-ca\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.409621 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/abc48bbe-a029-4324-b944-54e83092b7c2-trusted-ca-bundle\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.411192 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/abc48bbe-a029-4324-b944-54e83092b7c2-console-oauth-config\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.419962 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/abc48bbe-a029-4324-b944-54e83092b7c2-console-serving-cert\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.422702 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gqvn\" (UniqueName: \"kubernetes.io/projected/abc48bbe-a029-4324-b944-54e83092b7c2-kube-api-access-9gqvn\") pod \"console-75d6767b4d-z8c8m\" (UID: \"abc48bbe-a029-4324-b944-54e83092b7c2\") " pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.506429 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aea9659b-8952-4f71-a8d5-970bfd9fffe0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-kwnzx\" (UID: \"aea9659b-8952-4f71-a8d5-970bfd9fffe0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.515144 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aea9659b-8952-4f71-a8d5-970bfd9fffe0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-kwnzx\" (UID: \"aea9659b-8952-4f71-a8d5-970bfd9fffe0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.535074 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.597149 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh"] Dec 08 19:35:21 crc kubenswrapper[4706]: W1208 19:35:21.603783 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b6da606_8fe3_4228_81f7_533d8155c6f0.slice/crio-272ee599dee26a835381bbcdd3c85dbe3ac2409feee06f4d0a9905d67e9561ba WatchSource:0}: Error finding container 272ee599dee26a835381bbcdd3c85dbe3ac2409feee06f4d0a9905d67e9561ba: Status 404 returned error can't find the container with id 272ee599dee26a835381bbcdd3c85dbe3ac2409feee06f4d0a9905d67e9561ba Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.620528 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx"] Dec 08 19:35:21 crc kubenswrapper[4706]: W1208 19:35:21.623876 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05ea3b8c_286e_48e7_b232_ac550fcf165e.slice/crio-b58079c62aad766ecdf8f376e238ae63f64f1acf0c253acac53cdac3f002b59f WatchSource:0}: Error finding container b58079c62aad766ecdf8f376e238ae63f64f1acf0c253acac53cdac3f002b59f: Status 404 returned error can't find the container with id b58079c62aad766ecdf8f376e238ae63f64f1acf0c253acac53cdac3f002b59f Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.711142 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.788446 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-zqns4" event={"ID":"1287503f-533b-4f53-b9b2-6792ad891790","Type":"ContainerStarted","Data":"ddd26b2cd356b5b41a36f0958bc87fd07598ebfc95cb58656288ce2ea8b1eacd"} Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.791055 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" event={"ID":"0b6da606-8fe3-4228-81f7-533d8155c6f0","Type":"ContainerStarted","Data":"272ee599dee26a835381bbcdd3c85dbe3ac2409feee06f4d0a9905d67e9561ba"} Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.792489 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" event={"ID":"05ea3b8c-286e-48e7-b232-ac550fcf165e","Type":"ContainerStarted","Data":"b58079c62aad766ecdf8f376e238ae63f64f1acf0c253acac53cdac3f002b59f"} Dec 08 19:35:21 crc kubenswrapper[4706]: I1208 19:35:21.804105 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75d6767b4d-z8c8m"] Dec 08 19:35:21 crc kubenswrapper[4706]: W1208 19:35:21.819934 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabc48bbe_a029_4324_b944_54e83092b7c2.slice/crio-b4b0daf48adfe4816e5c765cfa2505c3277eb8af7ab483382d6cc5c4dc4a9a12 WatchSource:0}: Error finding container b4b0daf48adfe4816e5c765cfa2505c3277eb8af7ab483382d6cc5c4dc4a9a12: Status 404 returned error can't find the container with id b4b0daf48adfe4816e5c765cfa2505c3277eb8af7ab483382d6cc5c4dc4a9a12 Dec 08 19:35:22 crc kubenswrapper[4706]: I1208 19:35:22.136970 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx"] Dec 08 19:35:22 crc kubenswrapper[4706]: I1208 19:35:22.802429 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" event={"ID":"aea9659b-8952-4f71-a8d5-970bfd9fffe0","Type":"ContainerStarted","Data":"bbbe48d374406db26b0c8d945e248959a0158b0da25efb548352618fee223ae0"} Dec 08 19:35:22 crc kubenswrapper[4706]: I1208 19:35:22.803692 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75d6767b4d-z8c8m" event={"ID":"abc48bbe-a029-4324-b944-54e83092b7c2","Type":"ContainerStarted","Data":"9b79ada61cc4cdee659d19fe59ca341454517ecb71a5fcce763040430ab2b2c5"} Dec 08 19:35:22 crc kubenswrapper[4706]: I1208 19:35:22.803736 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75d6767b4d-z8c8m" event={"ID":"abc48bbe-a029-4324-b944-54e83092b7c2","Type":"ContainerStarted","Data":"b4b0daf48adfe4816e5c765cfa2505c3277eb8af7ab483382d6cc5c4dc4a9a12"} Dec 08 19:35:22 crc kubenswrapper[4706]: I1208 19:35:22.863198 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-75d6767b4d-z8c8m" podStartSLOduration=1.8631701280000001 podStartE2EDuration="1.863170128s" podCreationTimestamp="2025-12-08 19:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:35:22.829328471 +0000 UTC m=+805.471529494" watchObservedRunningTime="2025-12-08 19:35:22.863170128 +0000 UTC m=+805.505371131" Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.826807 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-zqns4" event={"ID":"1287503f-533b-4f53-b9b2-6792ad891790","Type":"ContainerStarted","Data":"741b333cdc64d3040e4aa6c48261cdd14eeb9833e2e0e6c2818fe47a6035b64e"} Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.827120 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.830014 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" event={"ID":"aea9659b-8952-4f71-a8d5-970bfd9fffe0","Type":"ContainerStarted","Data":"99d26459ac5056b343f6cf6589fa57441c678a44f6ff28867d11524cd258685c"} Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.830085 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.831778 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" event={"ID":"0b6da606-8fe3-4228-81f7-533d8155c6f0","Type":"ContainerStarted","Data":"0be954b319cf12c7f674547ab385fe2919a02451b361a8e75dbddfa26138a643"} Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.834511 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" event={"ID":"05ea3b8c-286e-48e7-b232-ac550fcf165e","Type":"ContainerStarted","Data":"cb42ea5c2c9c250cca0f370e93146364b60e4a55d41547755c99b08e211ca9f3"} Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.854138 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-zqns4" podStartSLOduration=1.910338726 podStartE2EDuration="5.854109664s" podCreationTimestamp="2025-12-08 19:35:20 +0000 UTC" firstStartedPulling="2025-12-08 19:35:21.234851298 +0000 UTC m=+803.877052301" lastFinishedPulling="2025-12-08 19:35:25.178622236 +0000 UTC m=+807.820823239" observedRunningTime="2025-12-08 19:35:25.850643136 +0000 UTC m=+808.492844149" watchObservedRunningTime="2025-12-08 19:35:25.854109664 +0000 UTC m=+808.496310667" Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.877345 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-224bh" podStartSLOduration=2.358236355 podStartE2EDuration="5.87731862s" podCreationTimestamp="2025-12-08 19:35:20 +0000 UTC" firstStartedPulling="2025-12-08 19:35:21.606888191 +0000 UTC m=+804.249089194" lastFinishedPulling="2025-12-08 19:35:25.125970456 +0000 UTC m=+807.768171459" observedRunningTime="2025-12-08 19:35:25.868576713 +0000 UTC m=+808.510777736" watchObservedRunningTime="2025-12-08 19:35:25.87731862 +0000 UTC m=+808.519519623" Dec 08 19:35:25 crc kubenswrapper[4706]: I1208 19:35:25.894311 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" podStartSLOduration=2.907429289 podStartE2EDuration="5.894250659s" podCreationTimestamp="2025-12-08 19:35:20 +0000 UTC" firstStartedPulling="2025-12-08 19:35:22.147603316 +0000 UTC m=+804.789804319" lastFinishedPulling="2025-12-08 19:35:25.134424686 +0000 UTC m=+807.776625689" observedRunningTime="2025-12-08 19:35:25.889156265 +0000 UTC m=+808.531357298" watchObservedRunningTime="2025-12-08 19:35:25.894250659 +0000 UTC m=+808.536451662" Dec 08 19:35:28 crc kubenswrapper[4706]: I1208 19:35:28.760025 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:28 crc kubenswrapper[4706]: I1208 19:35:28.810748 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:28 crc kubenswrapper[4706]: I1208 19:35:28.853152 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" event={"ID":"05ea3b8c-286e-48e7-b232-ac550fcf165e","Type":"ContainerStarted","Data":"2664aa53388c851b07ad4dcae90a459ec061f4e22531a470c46dd4cc62cee503"} Dec 08 19:35:28 crc kubenswrapper[4706]: I1208 19:35:28.875322 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-llfnx" podStartSLOduration=2.48358664 podStartE2EDuration="8.875295085s" podCreationTimestamp="2025-12-08 19:35:20 +0000 UTC" firstStartedPulling="2025-12-08 19:35:21.628339878 +0000 UTC m=+804.270540881" lastFinishedPulling="2025-12-08 19:35:28.020048323 +0000 UTC m=+810.662249326" observedRunningTime="2025-12-08 19:35:28.872615569 +0000 UTC m=+811.514816572" watchObservedRunningTime="2025-12-08 19:35:28.875295085 +0000 UTC m=+811.517496088" Dec 08 19:35:28 crc kubenswrapper[4706]: I1208 19:35:28.990646 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kxqlk"] Dec 08 19:35:29 crc kubenswrapper[4706]: I1208 19:35:29.859951 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kxqlk" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="registry-server" containerID="cri-o://4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a" gracePeriod=2 Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.258195 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.366914 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hxnb\" (UniqueName: \"kubernetes.io/projected/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-kube-api-access-5hxnb\") pod \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.367022 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-utilities\") pod \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.367126 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-catalog-content\") pod \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\" (UID: \"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6\") " Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.369463 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-utilities" (OuterVolumeSpecName: "utilities") pod "c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" (UID: "c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.377162 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-kube-api-access-5hxnb" (OuterVolumeSpecName: "kube-api-access-5hxnb") pod "c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" (UID: "c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6"). InnerVolumeSpecName "kube-api-access-5hxnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.468834 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hxnb\" (UniqueName: \"kubernetes.io/projected/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-kube-api-access-5hxnb\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.468884 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.470673 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" (UID: "c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.570741 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.870664 4706 generic.go:334] "Generic (PLEG): container finished" podID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerID="4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a" exitCode=0 Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.870725 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxqlk" event={"ID":"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6","Type":"ContainerDied","Data":"4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a"} Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.870758 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxqlk" event={"ID":"c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6","Type":"ContainerDied","Data":"b0fdf15898598676b7256f61ba0e2b7472472f1985660fcf41b41760bda0a192"} Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.870769 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxqlk" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.870778 4706 scope.go:117] "RemoveContainer" containerID="4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.892701 4706 scope.go:117] "RemoveContainer" containerID="2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.901611 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kxqlk"] Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.914232 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kxqlk"] Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.927487 4706 scope.go:117] "RemoveContainer" containerID="0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.958921 4706 scope.go:117] "RemoveContainer" containerID="4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a" Dec 08 19:35:30 crc kubenswrapper[4706]: E1208 19:35:30.959587 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a\": container with ID starting with 4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a not found: ID does not exist" containerID="4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.959700 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a"} err="failed to get container status \"4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a\": rpc error: code = NotFound desc = could not find container \"4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a\": container with ID starting with 4251620a9327a2be104639734f2501664861bc357a93b0ab1d338098e2d1b93a not found: ID does not exist" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.959740 4706 scope.go:117] "RemoveContainer" containerID="2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193" Dec 08 19:35:30 crc kubenswrapper[4706]: E1208 19:35:30.960144 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193\": container with ID starting with 2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193 not found: ID does not exist" containerID="2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.960191 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193"} err="failed to get container status \"2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193\": rpc error: code = NotFound desc = could not find container \"2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193\": container with ID starting with 2d82259aa923f4d0c46a13b27ea03fe251dc422ed1c341cf55d79a49a409d193 not found: ID does not exist" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.962003 4706 scope.go:117] "RemoveContainer" containerID="0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af" Dec 08 19:35:30 crc kubenswrapper[4706]: E1208 19:35:30.962475 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af\": container with ID starting with 0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af not found: ID does not exist" containerID="0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af" Dec 08 19:35:30 crc kubenswrapper[4706]: I1208 19:35:30.962547 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af"} err="failed to get container status \"0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af\": rpc error: code = NotFound desc = could not find container \"0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af\": container with ID starting with 0f095f327105227cbb583532c91dd42d0e803836f3f1f606ba547f449ba5b4af not found: ID does not exist" Dec 08 19:35:31 crc kubenswrapper[4706]: I1208 19:35:31.202076 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-zqns4" Dec 08 19:35:31 crc kubenswrapper[4706]: I1208 19:35:31.536753 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:31 crc kubenswrapper[4706]: I1208 19:35:31.536812 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:31 crc kubenswrapper[4706]: I1208 19:35:31.549402 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:31 crc kubenswrapper[4706]: I1208 19:35:31.617797 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" path="/var/lib/kubelet/pods/c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6/volumes" Dec 08 19:35:31 crc kubenswrapper[4706]: I1208 19:35:31.885319 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-75d6767b4d-z8c8m" Dec 08 19:35:31 crc kubenswrapper[4706]: I1208 19:35:31.993377 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zgngm"] Dec 08 19:35:35 crc kubenswrapper[4706]: I1208 19:35:35.836796 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:35:35 crc kubenswrapper[4706]: I1208 19:35:35.837412 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:35:41 crc kubenswrapper[4706]: I1208 19:35:41.717947 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kwnzx" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.198768 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf"] Dec 08 19:35:56 crc kubenswrapper[4706]: E1208 19:35:56.199825 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="extract-content" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.199844 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="extract-content" Dec 08 19:35:56 crc kubenswrapper[4706]: E1208 19:35:56.199857 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="extract-utilities" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.199864 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="extract-utilities" Dec 08 19:35:56 crc kubenswrapper[4706]: E1208 19:35:56.199874 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="registry-server" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.199882 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="registry-server" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.200029 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9d60181-6ac2-441f-8ca4-bf7c3a46ddd6" containerName="registry-server" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.201163 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.203384 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.207602 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf"] Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.258298 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.258562 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.258711 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mmn2\" (UniqueName: \"kubernetes.io/projected/07e73e83-48b7-43aa-b638-8c5898173956-kube-api-access-2mmn2\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.359747 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.359870 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.359933 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mmn2\" (UniqueName: \"kubernetes.io/projected/07e73e83-48b7-43aa-b638-8c5898173956-kube-api-access-2mmn2\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.360411 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.360418 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.379992 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mmn2\" (UniqueName: \"kubernetes.io/projected/07e73e83-48b7-43aa-b638-8c5898173956-kube-api-access-2mmn2\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.518109 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:35:56 crc kubenswrapper[4706]: I1208 19:35:56.775014 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf"] Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.042319 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-zgngm" podUID="3f1e0eea-50be-4dd8-a504-dcdd049ae255" containerName="console" containerID="cri-o://3da593dd8d025eec1bab4fbb886c67739cf46dba49a7488da0684e80e6385c29" gracePeriod=15 Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.092382 4706 generic.go:334] "Generic (PLEG): container finished" podID="07e73e83-48b7-43aa-b638-8c5898173956" containerID="3e8950a1212f7d0e7ad466b3f6b5eea2cb1d1738546dda15c508ac3a8c559aa1" exitCode=0 Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.092438 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" event={"ID":"07e73e83-48b7-43aa-b638-8c5898173956","Type":"ContainerDied","Data":"3e8950a1212f7d0e7ad466b3f6b5eea2cb1d1738546dda15c508ac3a8c559aa1"} Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.092472 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" event={"ID":"07e73e83-48b7-43aa-b638-8c5898173956","Type":"ContainerStarted","Data":"32ad925d3d023c1cfccd2762c4b0b40a634f37017672eaf8e0013e2768430693"} Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.414608 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zgngm_3f1e0eea-50be-4dd8-a504-dcdd049ae255/console/0.log" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.415002 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.592994 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-oauth-serving-cert\") pod \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.593409 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-service-ca\") pod \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.593462 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-trusted-ca-bundle\") pod \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.593504 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-config\") pod \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.593530 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-serving-cert\") pod \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.593541 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "3f1e0eea-50be-4dd8-a504-dcdd049ae255" (UID: "3f1e0eea-50be-4dd8-a504-dcdd049ae255"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.593581 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-oauth-config\") pod \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.593611 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgbw6\" (UniqueName: \"kubernetes.io/projected/3f1e0eea-50be-4dd8-a504-dcdd049ae255-kube-api-access-wgbw6\") pod \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\" (UID: \"3f1e0eea-50be-4dd8-a504-dcdd049ae255\") " Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.593961 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-service-ca" (OuterVolumeSpecName: "service-ca") pod "3f1e0eea-50be-4dd8-a504-dcdd049ae255" (UID: "3f1e0eea-50be-4dd8-a504-dcdd049ae255"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.594145 4706 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.594312 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-config" (OuterVolumeSpecName: "console-config") pod "3f1e0eea-50be-4dd8-a504-dcdd049ae255" (UID: "3f1e0eea-50be-4dd8-a504-dcdd049ae255"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.594414 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "3f1e0eea-50be-4dd8-a504-dcdd049ae255" (UID: "3f1e0eea-50be-4dd8-a504-dcdd049ae255"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.600081 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f1e0eea-50be-4dd8-a504-dcdd049ae255-kube-api-access-wgbw6" (OuterVolumeSpecName: "kube-api-access-wgbw6") pod "3f1e0eea-50be-4dd8-a504-dcdd049ae255" (UID: "3f1e0eea-50be-4dd8-a504-dcdd049ae255"). InnerVolumeSpecName "kube-api-access-wgbw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.600156 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "3f1e0eea-50be-4dd8-a504-dcdd049ae255" (UID: "3f1e0eea-50be-4dd8-a504-dcdd049ae255"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.600515 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "3f1e0eea-50be-4dd8-a504-dcdd049ae255" (UID: "3f1e0eea-50be-4dd8-a504-dcdd049ae255"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.694736 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgbw6\" (UniqueName: \"kubernetes.io/projected/3f1e0eea-50be-4dd8-a504-dcdd049ae255-kube-api-access-wgbw6\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.694767 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.694778 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.694787 4706 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.694796 4706 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.694803 4706 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3f1e0eea-50be-4dd8-a504-dcdd049ae255-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:35:57 crc kubenswrapper[4706]: I1208 19:35:57.936988 4706 scope.go:117] "RemoveContainer" containerID="3da593dd8d025eec1bab4fbb886c67739cf46dba49a7488da0684e80e6385c29" Dec 08 19:35:58 crc kubenswrapper[4706]: I1208 19:35:58.097226 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zgngm" event={"ID":"3f1e0eea-50be-4dd8-a504-dcdd049ae255","Type":"ContainerDied","Data":"3da593dd8d025eec1bab4fbb886c67739cf46dba49a7488da0684e80e6385c29"} Dec 08 19:35:58 crc kubenswrapper[4706]: I1208 19:35:58.097654 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zgngm" event={"ID":"3f1e0eea-50be-4dd8-a504-dcdd049ae255","Type":"ContainerDied","Data":"9c19b71f75c4ea9c54aa8af463b5b88662d8075b5ebbb11681b338e52223667b"} Dec 08 19:35:58 crc kubenswrapper[4706]: I1208 19:35:58.097261 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zgngm" Dec 08 19:35:58 crc kubenswrapper[4706]: I1208 19:35:58.143694 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zgngm"] Dec 08 19:35:58 crc kubenswrapper[4706]: I1208 19:35:58.151546 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-zgngm"] Dec 08 19:35:59 crc kubenswrapper[4706]: I1208 19:35:59.106552 4706 generic.go:334] "Generic (PLEG): container finished" podID="07e73e83-48b7-43aa-b638-8c5898173956" containerID="8744da2df2f44901e93bc1677df154ad675ff73e77bed2cc71557a51f8f314fb" exitCode=0 Dec 08 19:35:59 crc kubenswrapper[4706]: I1208 19:35:59.106605 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" event={"ID":"07e73e83-48b7-43aa-b638-8c5898173956","Type":"ContainerDied","Data":"8744da2df2f44901e93bc1677df154ad675ff73e77bed2cc71557a51f8f314fb"} Dec 08 19:35:59 crc kubenswrapper[4706]: I1208 19:35:59.616330 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f1e0eea-50be-4dd8-a504-dcdd049ae255" path="/var/lib/kubelet/pods/3f1e0eea-50be-4dd8-a504-dcdd049ae255/volumes" Dec 08 19:36:00 crc kubenswrapper[4706]: I1208 19:36:00.115750 4706 generic.go:334] "Generic (PLEG): container finished" podID="07e73e83-48b7-43aa-b638-8c5898173956" containerID="afc6478273e0025b0109fc9fcd37182f68caf404d17dea5c3fa33f9ebd31fb5a" exitCode=0 Dec 08 19:36:00 crc kubenswrapper[4706]: I1208 19:36:00.115800 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" event={"ID":"07e73e83-48b7-43aa-b638-8c5898173956","Type":"ContainerDied","Data":"afc6478273e0025b0109fc9fcd37182f68caf404d17dea5c3fa33f9ebd31fb5a"} Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.356035 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.544198 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mmn2\" (UniqueName: \"kubernetes.io/projected/07e73e83-48b7-43aa-b638-8c5898173956-kube-api-access-2mmn2\") pod \"07e73e83-48b7-43aa-b638-8c5898173956\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.544292 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-util\") pod \"07e73e83-48b7-43aa-b638-8c5898173956\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.544413 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-bundle\") pod \"07e73e83-48b7-43aa-b638-8c5898173956\" (UID: \"07e73e83-48b7-43aa-b638-8c5898173956\") " Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.545841 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-bundle" (OuterVolumeSpecName: "bundle") pod "07e73e83-48b7-43aa-b638-8c5898173956" (UID: "07e73e83-48b7-43aa-b638-8c5898173956"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.551528 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07e73e83-48b7-43aa-b638-8c5898173956-kube-api-access-2mmn2" (OuterVolumeSpecName: "kube-api-access-2mmn2") pod "07e73e83-48b7-43aa-b638-8c5898173956" (UID: "07e73e83-48b7-43aa-b638-8c5898173956"). InnerVolumeSpecName "kube-api-access-2mmn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.559480 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-util" (OuterVolumeSpecName: "util") pod "07e73e83-48b7-43aa-b638-8c5898173956" (UID: "07e73e83-48b7-43aa-b638-8c5898173956"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.646487 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-util\") on node \"crc\" DevicePath \"\"" Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.646524 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07e73e83-48b7-43aa-b638-8c5898173956-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:36:01 crc kubenswrapper[4706]: I1208 19:36:01.646534 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mmn2\" (UniqueName: \"kubernetes.io/projected/07e73e83-48b7-43aa-b638-8c5898173956-kube-api-access-2mmn2\") on node \"crc\" DevicePath \"\"" Dec 08 19:36:02 crc kubenswrapper[4706]: I1208 19:36:02.149598 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" event={"ID":"07e73e83-48b7-43aa-b638-8c5898173956","Type":"ContainerDied","Data":"32ad925d3d023c1cfccd2762c4b0b40a634f37017672eaf8e0013e2768430693"} Dec 08 19:36:02 crc kubenswrapper[4706]: I1208 19:36:02.149681 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32ad925d3d023c1cfccd2762c4b0b40a634f37017672eaf8e0013e2768430693" Dec 08 19:36:02 crc kubenswrapper[4706]: I1208 19:36:02.149773 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf" Dec 08 19:36:05 crc kubenswrapper[4706]: I1208 19:36:05.835937 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:36:05 crc kubenswrapper[4706]: I1208 19:36:05.836649 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:36:05 crc kubenswrapper[4706]: I1208 19:36:05.836699 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:36:05 crc kubenswrapper[4706]: I1208 19:36:05.837414 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d44a6096d27dcbd84b0ab9d9f054b00bc8d565b62f431d2f781917837a86f52"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:36:05 crc kubenswrapper[4706]: I1208 19:36:05.837470 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://7d44a6096d27dcbd84b0ab9d9f054b00bc8d565b62f431d2f781917837a86f52" gracePeriod=600 Dec 08 19:36:06 crc kubenswrapper[4706]: I1208 19:36:06.176487 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="7d44a6096d27dcbd84b0ab9d9f054b00bc8d565b62f431d2f781917837a86f52" exitCode=0 Dec 08 19:36:06 crc kubenswrapper[4706]: I1208 19:36:06.176538 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"7d44a6096d27dcbd84b0ab9d9f054b00bc8d565b62f431d2f781917837a86f52"} Dec 08 19:36:06 crc kubenswrapper[4706]: I1208 19:36:06.176581 4706 scope.go:117] "RemoveContainer" containerID="00eeebf4bd5f58c323dc15ecfd6b82b6e5fadacd75c5f1d8bddcf05b29f53f47" Dec 08 19:36:07 crc kubenswrapper[4706]: I1208 19:36:07.185142 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"44273cc3a85c65641d37bbdee9218aa6e6f7747d6e135429b54fe362e247061f"} Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.962969 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj"] Dec 08 19:36:10 crc kubenswrapper[4706]: E1208 19:36:10.964017 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1e0eea-50be-4dd8-a504-dcdd049ae255" containerName="console" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.964036 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1e0eea-50be-4dd8-a504-dcdd049ae255" containerName="console" Dec 08 19:36:10 crc kubenswrapper[4706]: E1208 19:36:10.964053 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e73e83-48b7-43aa-b638-8c5898173956" containerName="util" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.964060 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e73e83-48b7-43aa-b638-8c5898173956" containerName="util" Dec 08 19:36:10 crc kubenswrapper[4706]: E1208 19:36:10.964070 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e73e83-48b7-43aa-b638-8c5898173956" containerName="extract" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.964081 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e73e83-48b7-43aa-b638-8c5898173956" containerName="extract" Dec 08 19:36:10 crc kubenswrapper[4706]: E1208 19:36:10.964097 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e73e83-48b7-43aa-b638-8c5898173956" containerName="pull" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.964104 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e73e83-48b7-43aa-b638-8c5898173956" containerName="pull" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.964232 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="07e73e83-48b7-43aa-b638-8c5898173956" containerName="extract" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.964251 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f1e0eea-50be-4dd8-a504-dcdd049ae255" containerName="console" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.964837 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.968595 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.970123 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.970698 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.971164 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-lghpj" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.971501 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 08 19:36:10 crc kubenswrapper[4706]: I1208 19:36:10.991475 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj"] Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.079807 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fdc00423-9550-4ff3-84d9-78eeb50e29ea-webhook-cert\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.079953 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gfhz\" (UniqueName: \"kubernetes.io/projected/fdc00423-9550-4ff3-84d9-78eeb50e29ea-kube-api-access-7gfhz\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.079987 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fdc00423-9550-4ff3-84d9-78eeb50e29ea-apiservice-cert\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.181901 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fdc00423-9550-4ff3-84d9-78eeb50e29ea-webhook-cert\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.183671 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gfhz\" (UniqueName: \"kubernetes.io/projected/fdc00423-9550-4ff3-84d9-78eeb50e29ea-kube-api-access-7gfhz\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.183701 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fdc00423-9550-4ff3-84d9-78eeb50e29ea-apiservice-cert\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.188976 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fdc00423-9550-4ff3-84d9-78eeb50e29ea-webhook-cert\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.197975 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fdc00423-9550-4ff3-84d9-78eeb50e29ea-apiservice-cert\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.206056 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gfhz\" (UniqueName: \"kubernetes.io/projected/fdc00423-9550-4ff3-84d9-78eeb50e29ea-kube-api-access-7gfhz\") pod \"metallb-operator-controller-manager-65c6ff64d8-bdjmj\" (UID: \"fdc00423-9550-4ff3-84d9-78eeb50e29ea\") " pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.281142 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.282074 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-75457555df-ll7rx"] Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.282889 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.287880 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.287968 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.287977 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pxtz6" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.317360 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75457555df-ll7rx"] Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.385994 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b87g\" (UniqueName: \"kubernetes.io/projected/4ef64783-e4c8-4c02-884f-3dba766387b8-kube-api-access-4b87g\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.386078 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4ef64783-e4c8-4c02-884f-3dba766387b8-apiservice-cert\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.386118 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4ef64783-e4c8-4c02-884f-3dba766387b8-webhook-cert\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.488186 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b87g\" (UniqueName: \"kubernetes.io/projected/4ef64783-e4c8-4c02-884f-3dba766387b8-kube-api-access-4b87g\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.488293 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4ef64783-e4c8-4c02-884f-3dba766387b8-apiservice-cert\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.488345 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4ef64783-e4c8-4c02-884f-3dba766387b8-webhook-cert\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.493647 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4ef64783-e4c8-4c02-884f-3dba766387b8-apiservice-cert\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.502529 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4ef64783-e4c8-4c02-884f-3dba766387b8-webhook-cert\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.550451 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b87g\" (UniqueName: \"kubernetes.io/projected/4ef64783-e4c8-4c02-884f-3dba766387b8-kube-api-access-4b87g\") pod \"metallb-operator-webhook-server-75457555df-ll7rx\" (UID: \"4ef64783-e4c8-4c02-884f-3dba766387b8\") " pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.650215 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:11 crc kubenswrapper[4706]: W1208 19:36:11.886939 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdc00423_9550_4ff3_84d9_78eeb50e29ea.slice/crio-dd7e05d1bf72f42c37ecdae8b9074eaea38c577cbc3648aea112ae4c4f485b38 WatchSource:0}: Error finding container dd7e05d1bf72f42c37ecdae8b9074eaea38c577cbc3648aea112ae4c4f485b38: Status 404 returned error can't find the container with id dd7e05d1bf72f42c37ecdae8b9074eaea38c577cbc3648aea112ae4c4f485b38 Dec 08 19:36:11 crc kubenswrapper[4706]: I1208 19:36:11.887967 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj"] Dec 08 19:36:12 crc kubenswrapper[4706]: I1208 19:36:12.083504 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75457555df-ll7rx"] Dec 08 19:36:12 crc kubenswrapper[4706]: W1208 19:36:12.091058 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ef64783_e4c8_4c02_884f_3dba766387b8.slice/crio-6a0a1be7cf2ba5d060258fa8b1612c6a17f526a8843503104fb794ce6e1eec3e WatchSource:0}: Error finding container 6a0a1be7cf2ba5d060258fa8b1612c6a17f526a8843503104fb794ce6e1eec3e: Status 404 returned error can't find the container with id 6a0a1be7cf2ba5d060258fa8b1612c6a17f526a8843503104fb794ce6e1eec3e Dec 08 19:36:12 crc kubenswrapper[4706]: I1208 19:36:12.225018 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" event={"ID":"fdc00423-9550-4ff3-84d9-78eeb50e29ea","Type":"ContainerStarted","Data":"dd7e05d1bf72f42c37ecdae8b9074eaea38c577cbc3648aea112ae4c4f485b38"} Dec 08 19:36:12 crc kubenswrapper[4706]: I1208 19:36:12.226520 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" event={"ID":"4ef64783-e4c8-4c02-884f-3dba766387b8","Type":"ContainerStarted","Data":"6a0a1be7cf2ba5d060258fa8b1612c6a17f526a8843503104fb794ce6e1eec3e"} Dec 08 19:36:17 crc kubenswrapper[4706]: I1208 19:36:17.287529 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" event={"ID":"4ef64783-e4c8-4c02-884f-3dba766387b8","Type":"ContainerStarted","Data":"6022632ec040601d363e6fe1bab4094be99b4ae2986f7abcaec92b5179d607b8"} Dec 08 19:36:17 crc kubenswrapper[4706]: I1208 19:36:17.288206 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:17 crc kubenswrapper[4706]: I1208 19:36:17.289824 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" event={"ID":"fdc00423-9550-4ff3-84d9-78eeb50e29ea","Type":"ContainerStarted","Data":"33737dd6042ea334a88bd986bde628dce80c63f135133ec9cc6c37f6bb33e9b2"} Dec 08 19:36:17 crc kubenswrapper[4706]: I1208 19:36:17.289954 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:17 crc kubenswrapper[4706]: I1208 19:36:17.308062 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" podStartSLOduration=1.519708216 podStartE2EDuration="6.308044806s" podCreationTimestamp="2025-12-08 19:36:11 +0000 UTC" firstStartedPulling="2025-12-08 19:36:12.093727461 +0000 UTC m=+854.735928464" lastFinishedPulling="2025-12-08 19:36:16.882064051 +0000 UTC m=+859.524265054" observedRunningTime="2025-12-08 19:36:17.305115813 +0000 UTC m=+859.947316816" watchObservedRunningTime="2025-12-08 19:36:17.308044806 +0000 UTC m=+859.950245809" Dec 08 19:36:17 crc kubenswrapper[4706]: I1208 19:36:17.333583 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" podStartSLOduration=2.365787223 podStartE2EDuration="7.333551287s" podCreationTimestamp="2025-12-08 19:36:10 +0000 UTC" firstStartedPulling="2025-12-08 19:36:11.889150836 +0000 UTC m=+854.531351839" lastFinishedPulling="2025-12-08 19:36:16.8569149 +0000 UTC m=+859.499115903" observedRunningTime="2025-12-08 19:36:17.328978338 +0000 UTC m=+859.971179341" watchObservedRunningTime="2025-12-08 19:36:17.333551287 +0000 UTC m=+859.975752290" Dec 08 19:36:31 crc kubenswrapper[4706]: I1208 19:36:31.656386 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-75457555df-ll7rx" Dec 08 19:36:51 crc kubenswrapper[4706]: I1208 19:36:51.284822 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-65c6ff64d8-bdjmj" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.057323 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jtzsv"] Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.060502 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.062169 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz"] Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.063065 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.064812 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.065077 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.067019 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.067204 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-57fq2" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.085462 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz"] Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.168561 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-49kgz"] Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.169823 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.172529 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-96rdm" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.172809 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.172954 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.176917 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.192343 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-9k84w"] Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.193896 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.199947 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.214375 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-9k84w"] Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.242119 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/257f3480-2853-4319-b4c2-7eb65e6bb097-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mdfmz\" (UID: \"257f3480-2853-4319-b4c2-7eb65e6bb097\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.242474 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-reloader\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.242559 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1af879c2-1138-4ca6-92d5-013c50cb4ef2-metrics-certs\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.242625 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-sockets\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.242691 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-conf\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.242805 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-metrics\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.242896 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph2r4\" (UniqueName: \"kubernetes.io/projected/257f3480-2853-4319-b4c2-7eb65e6bb097-kube-api-access-ph2r4\") pod \"frr-k8s-webhook-server-7fcb986d4-mdfmz\" (UID: \"257f3480-2853-4319-b4c2-7eb65e6bb097\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.242977 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk5qv\" (UniqueName: \"kubernetes.io/projected/1af879c2-1138-4ca6-92d5-013c50cb4ef2-kube-api-access-qk5qv\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.243054 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-startup\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345014 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345172 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-metrics\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345233 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1398dd3e-9aac-4f48-b927-1862ed71c544-metallb-excludel2\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345269 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph2r4\" (UniqueName: \"kubernetes.io/projected/257f3480-2853-4319-b4c2-7eb65e6bb097-kube-api-access-ph2r4\") pod \"frr-k8s-webhook-server-7fcb986d4-mdfmz\" (UID: \"257f3480-2853-4319-b4c2-7eb65e6bb097\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345318 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65z7h\" (UniqueName: \"kubernetes.io/projected/1398dd3e-9aac-4f48-b927-1862ed71c544-kube-api-access-65z7h\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345338 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-cert\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345364 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-metrics-certs\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345397 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk5qv\" (UniqueName: \"kubernetes.io/projected/1af879c2-1138-4ca6-92d5-013c50cb4ef2-kube-api-access-qk5qv\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345430 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-startup\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345459 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-metrics-certs\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345486 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zfs2\" (UniqueName: \"kubernetes.io/projected/00d01f1b-f2d8-467d-94a3-e9a73966d334-kube-api-access-7zfs2\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345515 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/257f3480-2853-4319-b4c2-7eb65e6bb097-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mdfmz\" (UID: \"257f3480-2853-4319-b4c2-7eb65e6bb097\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345548 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-reloader\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345570 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1af879c2-1138-4ca6-92d5-013c50cb4ef2-metrics-certs\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345593 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-sockets\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345614 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-conf\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.345624 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-metrics\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.346219 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-conf\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.346601 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-startup\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.346808 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-frr-sockets\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.347051 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1af879c2-1138-4ca6-92d5-013c50cb4ef2-reloader\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.357593 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1af879c2-1138-4ca6-92d5-013c50cb4ef2-metrics-certs\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.378921 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/257f3480-2853-4319-b4c2-7eb65e6bb097-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mdfmz\" (UID: \"257f3480-2853-4319-b4c2-7eb65e6bb097\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.381250 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph2r4\" (UniqueName: \"kubernetes.io/projected/257f3480-2853-4319-b4c2-7eb65e6bb097-kube-api-access-ph2r4\") pod \"frr-k8s-webhook-server-7fcb986d4-mdfmz\" (UID: \"257f3480-2853-4319-b4c2-7eb65e6bb097\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.384047 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk5qv\" (UniqueName: \"kubernetes.io/projected/1af879c2-1138-4ca6-92d5-013c50cb4ef2-kube-api-access-qk5qv\") pod \"frr-k8s-jtzsv\" (UID: \"1af879c2-1138-4ca6-92d5-013c50cb4ef2\") " pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.384491 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.446248 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.446357 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1398dd3e-9aac-4f48-b927-1862ed71c544-metallb-excludel2\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.446399 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65z7h\" (UniqueName: \"kubernetes.io/projected/1398dd3e-9aac-4f48-b927-1862ed71c544-kube-api-access-65z7h\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.446418 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-cert\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.446436 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-metrics-certs\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.446473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-metrics-certs\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.446496 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zfs2\" (UniqueName: \"kubernetes.io/projected/00d01f1b-f2d8-467d-94a3-e9a73966d334-kube-api-access-7zfs2\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: E1208 19:36:52.446554 4706 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 08 19:36:52 crc kubenswrapper[4706]: E1208 19:36:52.446651 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist podName:1398dd3e-9aac-4f48-b927-1862ed71c544 nodeName:}" failed. No retries permitted until 2025-12-08 19:36:52.946620451 +0000 UTC m=+895.588821634 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist") pod "speaker-49kgz" (UID: "1398dd3e-9aac-4f48-b927-1862ed71c544") : secret "metallb-memberlist" not found Dec 08 19:36:52 crc kubenswrapper[4706]: E1208 19:36:52.447033 4706 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 08 19:36:52 crc kubenswrapper[4706]: E1208 19:36:52.447097 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-metrics-certs podName:00d01f1b-f2d8-467d-94a3-e9a73966d334 nodeName:}" failed. No retries permitted until 2025-12-08 19:36:52.947079564 +0000 UTC m=+895.589280567 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-metrics-certs") pod "controller-f8648f98b-9k84w" (UID: "00d01f1b-f2d8-467d-94a3-e9a73966d334") : secret "controller-certs-secret" not found Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.447922 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1398dd3e-9aac-4f48-b927-1862ed71c544-metallb-excludel2\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.457153 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-cert\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.465972 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-metrics-certs\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.488571 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65z7h\" (UniqueName: \"kubernetes.io/projected/1398dd3e-9aac-4f48-b927-1862ed71c544-kube-api-access-65z7h\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.488900 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zfs2\" (UniqueName: \"kubernetes.io/projected/00d01f1b-f2d8-467d-94a3-e9a73966d334-kube-api-access-7zfs2\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.678083 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.951210 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz"] Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.953263 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-metrics-certs\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.953597 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:52 crc kubenswrapper[4706]: E1208 19:36:52.953813 4706 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 08 19:36:52 crc kubenswrapper[4706]: E1208 19:36:52.953884 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist podName:1398dd3e-9aac-4f48-b927-1862ed71c544 nodeName:}" failed. No retries permitted until 2025-12-08 19:36:53.953865563 +0000 UTC m=+896.596066566 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist") pod "speaker-49kgz" (UID: "1398dd3e-9aac-4f48-b927-1862ed71c544") : secret "metallb-memberlist" not found Dec 08 19:36:52 crc kubenswrapper[4706]: I1208 19:36:52.961765 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00d01f1b-f2d8-467d-94a3-e9a73966d334-metrics-certs\") pod \"controller-f8648f98b-9k84w\" (UID: \"00d01f1b-f2d8-467d-94a3-e9a73966d334\") " pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.049034 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dtffp"] Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.050983 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.055251 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-utilities\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.055554 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98dv6\" (UniqueName: \"kubernetes.io/projected/c8a77b1e-cb15-4263-91f2-a437dafb360e-kube-api-access-98dv6\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.055634 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-catalog-content\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.066611 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtffp"] Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.116495 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.157655 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-utilities\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.157770 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98dv6\" (UniqueName: \"kubernetes.io/projected/c8a77b1e-cb15-4263-91f2-a437dafb360e-kube-api-access-98dv6\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.157804 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-catalog-content\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.158604 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-utilities\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.158626 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-catalog-content\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.175161 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98dv6\" (UniqueName: \"kubernetes.io/projected/c8a77b1e-cb15-4263-91f2-a437dafb360e-kube-api-access-98dv6\") pod \"redhat-marketplace-dtffp\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.357430 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-9k84w"] Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.372949 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:36:53 crc kubenswrapper[4706]: W1208 19:36:53.376604 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00d01f1b_f2d8_467d_94a3_e9a73966d334.slice/crio-788afc8cc3faeeaf934a699834a09edef80f835e5ad46bf259288d2bf16e8530 WatchSource:0}: Error finding container 788afc8cc3faeeaf934a699834a09edef80f835e5ad46bf259288d2bf16e8530: Status 404 returned error can't find the container with id 788afc8cc3faeeaf934a699834a09edef80f835e5ad46bf259288d2bf16e8530 Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.534635 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" event={"ID":"257f3480-2853-4319-b4c2-7eb65e6bb097","Type":"ContainerStarted","Data":"abebbdcc1f27837a3bc646546d67bd0e538de4a39f4c08fa2455c3d1cfe70143"} Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.536071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-9k84w" event={"ID":"00d01f1b-f2d8-467d-94a3-e9a73966d334","Type":"ContainerStarted","Data":"788afc8cc3faeeaf934a699834a09edef80f835e5ad46bf259288d2bf16e8530"} Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.537728 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerStarted","Data":"268ff436900b74664b75e08da30f97200dffaccc04ab6e163480bd1b6d568d28"} Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.740439 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtffp"] Dec 08 19:36:53 crc kubenswrapper[4706]: W1208 19:36:53.747659 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8a77b1e_cb15_4263_91f2_a437dafb360e.slice/crio-73a27ad033cb13de11b8f0cde6856d0aedec9db5398e1e0ef26c67bd85b732a2 WatchSource:0}: Error finding container 73a27ad033cb13de11b8f0cde6856d0aedec9db5398e1e0ef26c67bd85b732a2: Status 404 returned error can't find the container with id 73a27ad033cb13de11b8f0cde6856d0aedec9db5398e1e0ef26c67bd85b732a2 Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.971845 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.979095 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1398dd3e-9aac-4f48-b927-1862ed71c544-memberlist\") pod \"speaker-49kgz\" (UID: \"1398dd3e-9aac-4f48-b927-1862ed71c544\") " pod="metallb-system/speaker-49kgz" Dec 08 19:36:53 crc kubenswrapper[4706]: I1208 19:36:53.989137 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-49kgz" Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.556854 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-9k84w" event={"ID":"00d01f1b-f2d8-467d-94a3-e9a73966d334","Type":"ContainerStarted","Data":"409db72007ceec5d3e0508319e978902d8ecff3a11049308c27b9d8c64dddc8b"} Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.557284 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-9k84w" event={"ID":"00d01f1b-f2d8-467d-94a3-e9a73966d334","Type":"ContainerStarted","Data":"05d29ffa443f741ac88d0d5ae9f3d233d61502bb63947b20d0176d71fd5ad255"} Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.558510 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.564140 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtffp" event={"ID":"c8a77b1e-cb15-4263-91f2-a437dafb360e","Type":"ContainerDied","Data":"f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf"} Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.563793 4706 generic.go:334] "Generic (PLEG): container finished" podID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerID="f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf" exitCode=0 Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.564668 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtffp" event={"ID":"c8a77b1e-cb15-4263-91f2-a437dafb360e","Type":"ContainerStarted","Data":"73a27ad033cb13de11b8f0cde6856d0aedec9db5398e1e0ef26c67bd85b732a2"} Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.566579 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-49kgz" event={"ID":"1398dd3e-9aac-4f48-b927-1862ed71c544","Type":"ContainerStarted","Data":"ab9d7b978518fbb729bb7766fde1d623df5fa173020b71f73c423721f034a1f5"} Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.566611 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-49kgz" event={"ID":"1398dd3e-9aac-4f48-b927-1862ed71c544","Type":"ContainerStarted","Data":"096dba44d1a58d2f1fa494e50a0b1f938f8e29bb6c211173dacb285577e72607"} Dec 08 19:36:54 crc kubenswrapper[4706]: I1208 19:36:54.582613 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-9k84w" podStartSLOduration=2.582583404 podStartE2EDuration="2.582583404s" podCreationTimestamp="2025-12-08 19:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:36:54.576523213 +0000 UTC m=+897.218724226" watchObservedRunningTime="2025-12-08 19:36:54.582583404 +0000 UTC m=+897.224784407" Dec 08 19:36:55 crc kubenswrapper[4706]: I1208 19:36:55.594119 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtffp" event={"ID":"c8a77b1e-cb15-4263-91f2-a437dafb360e","Type":"ContainerStarted","Data":"37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa"} Dec 08 19:36:55 crc kubenswrapper[4706]: I1208 19:36:55.622406 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-49kgz" event={"ID":"1398dd3e-9aac-4f48-b927-1862ed71c544","Type":"ContainerStarted","Data":"5660a7fd661849fd3b3b23b082fec7a1c7371dfd398d67cde82c284d7d225640"} Dec 08 19:36:55 crc kubenswrapper[4706]: I1208 19:36:55.632967 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-49kgz" podStartSLOduration=3.632942454 podStartE2EDuration="3.632942454s" podCreationTimestamp="2025-12-08 19:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:36:55.630387381 +0000 UTC m=+898.272588394" watchObservedRunningTime="2025-12-08 19:36:55.632942454 +0000 UTC m=+898.275143457" Dec 08 19:36:56 crc kubenswrapper[4706]: I1208 19:36:56.617520 4706 generic.go:334] "Generic (PLEG): container finished" podID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerID="37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa" exitCode=0 Dec 08 19:36:56 crc kubenswrapper[4706]: I1208 19:36:56.618950 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtffp" event={"ID":"c8a77b1e-cb15-4263-91f2-a437dafb360e","Type":"ContainerDied","Data":"37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa"} Dec 08 19:36:56 crc kubenswrapper[4706]: I1208 19:36:56.619371 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-49kgz" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.447583 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bsh55"] Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.449890 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.462439 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bsh55"] Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.562447 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gkwr\" (UniqueName: \"kubernetes.io/projected/a6913af1-1b8d-4d14-8b41-7f44443f88f6-kube-api-access-5gkwr\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.562615 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-utilities\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.562642 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-catalog-content\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.664698 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-catalog-content\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.664756 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-utilities\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.664796 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gkwr\" (UniqueName: \"kubernetes.io/projected/a6913af1-1b8d-4d14-8b41-7f44443f88f6-kube-api-access-5gkwr\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.665869 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-catalog-content\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.665890 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-utilities\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.699481 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gkwr\" (UniqueName: \"kubernetes.io/projected/a6913af1-1b8d-4d14-8b41-7f44443f88f6-kube-api-access-5gkwr\") pod \"certified-operators-bsh55\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:36:58 crc kubenswrapper[4706]: I1208 19:36:58.813124 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.426381 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bsh55"] Dec 08 19:37:02 crc kubenswrapper[4706]: W1208 19:37:02.433139 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6913af1_1b8d_4d14_8b41_7f44443f88f6.slice/crio-45623180e411d2b9c08346386e0b98e0a599816f86438f9e409f9ee0690da54b WatchSource:0}: Error finding container 45623180e411d2b9c08346386e0b98e0a599816f86438f9e409f9ee0690da54b: Status 404 returned error can't find the container with id 45623180e411d2b9c08346386e0b98e0a599816f86438f9e409f9ee0690da54b Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.665728 4706 generic.go:334] "Generic (PLEG): container finished" podID="1af879c2-1138-4ca6-92d5-013c50cb4ef2" containerID="91a73172e51a2ceb7b5db30aab158839427bedd68c3184182bf51fc571f2877f" exitCode=0 Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.665850 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerDied","Data":"91a73172e51a2ceb7b5db30aab158839427bedd68c3184182bf51fc571f2877f"} Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.668999 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtffp" event={"ID":"c8a77b1e-cb15-4263-91f2-a437dafb360e","Type":"ContainerStarted","Data":"5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608"} Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.671877 4706 generic.go:334] "Generic (PLEG): container finished" podID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerID="f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842" exitCode=0 Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.671957 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsh55" event={"ID":"a6913af1-1b8d-4d14-8b41-7f44443f88f6","Type":"ContainerDied","Data":"f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842"} Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.671987 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsh55" event={"ID":"a6913af1-1b8d-4d14-8b41-7f44443f88f6","Type":"ContainerStarted","Data":"45623180e411d2b9c08346386e0b98e0a599816f86438f9e409f9ee0690da54b"} Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.674302 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" event={"ID":"257f3480-2853-4319-b4c2-7eb65e6bb097","Type":"ContainerStarted","Data":"2249774cdc935276dcb5400452a0ab4af7ac344e94ae66f74b38ed1c0ebae15f"} Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.674743 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.732426 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" podStartSLOduration=1.716703473 podStartE2EDuration="10.732407331s" podCreationTimestamp="2025-12-08 19:36:52 +0000 UTC" firstStartedPulling="2025-12-08 19:36:52.956397505 +0000 UTC m=+895.598598528" lastFinishedPulling="2025-12-08 19:37:01.972101383 +0000 UTC m=+904.614302386" observedRunningTime="2025-12-08 19:37:02.727635806 +0000 UTC m=+905.369836809" watchObservedRunningTime="2025-12-08 19:37:02.732407331 +0000 UTC m=+905.374608334" Dec 08 19:37:02 crc kubenswrapper[4706]: I1208 19:37:02.756118 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dtffp" podStartSLOduration=2.447885281 podStartE2EDuration="9.7560844s" podCreationTimestamp="2025-12-08 19:36:53 +0000 UTC" firstStartedPulling="2025-12-08 19:36:54.565592964 +0000 UTC m=+897.207793967" lastFinishedPulling="2025-12-08 19:37:01.873792083 +0000 UTC m=+904.515993086" observedRunningTime="2025-12-08 19:37:02.751387488 +0000 UTC m=+905.393588501" watchObservedRunningTime="2025-12-08 19:37:02.7560844 +0000 UTC m=+905.398285403" Dec 08 19:37:03 crc kubenswrapper[4706]: I1208 19:37:03.123133 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-9k84w" Dec 08 19:37:03 crc kubenswrapper[4706]: I1208 19:37:03.373651 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:37:03 crc kubenswrapper[4706]: I1208 19:37:03.373967 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:37:03 crc kubenswrapper[4706]: I1208 19:37:03.681374 4706 generic.go:334] "Generic (PLEG): container finished" podID="1af879c2-1138-4ca6-92d5-013c50cb4ef2" containerID="05d1e722c15811586e59de3a4ff17373569b6aa0500255a9afe0f2dd0693dcdd" exitCode=0 Dec 08 19:37:03 crc kubenswrapper[4706]: I1208 19:37:03.681495 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerDied","Data":"05d1e722c15811586e59de3a4ff17373569b6aa0500255a9afe0f2dd0693dcdd"} Dec 08 19:37:04 crc kubenswrapper[4706]: I1208 19:37:04.417738 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-dtffp" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="registry-server" probeResult="failure" output=< Dec 08 19:37:04 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:37:04 crc kubenswrapper[4706]: > Dec 08 19:37:04 crc kubenswrapper[4706]: I1208 19:37:04.690948 4706 generic.go:334] "Generic (PLEG): container finished" podID="1af879c2-1138-4ca6-92d5-013c50cb4ef2" containerID="16da4e95037d81f3ed373148da796dcbef77321c19a5189cfd631fc8a50a5cc2" exitCode=0 Dec 08 19:37:04 crc kubenswrapper[4706]: I1208 19:37:04.691019 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerDied","Data":"16da4e95037d81f3ed373148da796dcbef77321c19a5189cfd631fc8a50a5cc2"} Dec 08 19:37:04 crc kubenswrapper[4706]: I1208 19:37:04.693030 4706 generic.go:334] "Generic (PLEG): container finished" podID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerID="df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6" exitCode=0 Dec 08 19:37:04 crc kubenswrapper[4706]: I1208 19:37:04.694544 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsh55" event={"ID":"a6913af1-1b8d-4d14-8b41-7f44443f88f6","Type":"ContainerDied","Data":"df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6"} Dec 08 19:37:05 crc kubenswrapper[4706]: I1208 19:37:05.704626 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerStarted","Data":"4cfb91b5acde6857d9d1b72c3ce01403d6eddf5aee42b7dbc94bba7c39cbf8dd"} Dec 08 19:37:05 crc kubenswrapper[4706]: I1208 19:37:05.705087 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerStarted","Data":"66de27e9588ef3c579f81fc715ec8634134e715dfc1e8dae6891f3fba29e9914"} Dec 08 19:37:05 crc kubenswrapper[4706]: I1208 19:37:05.705105 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerStarted","Data":"0bc52e07d8514b98131c6d564153c37dec8cb17d0e71076033ab582de8690d79"} Dec 08 19:37:05 crc kubenswrapper[4706]: I1208 19:37:05.705119 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerStarted","Data":"b58a4fe6197a6d0a083cc6b29657ab3c6ee80b7556d414950baeb41fa5e072c5"} Dec 08 19:37:05 crc kubenswrapper[4706]: I1208 19:37:05.705456 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerStarted","Data":"67e629c35d3223cf7eceddeec261cb2b1e0eb2fbd6a74ed199c09372d895b3ca"} Dec 08 19:37:05 crc kubenswrapper[4706]: I1208 19:37:05.710096 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsh55" event={"ID":"a6913af1-1b8d-4d14-8b41-7f44443f88f6","Type":"ContainerStarted","Data":"c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b"} Dec 08 19:37:05 crc kubenswrapper[4706]: I1208 19:37:05.729134 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bsh55" podStartSLOduration=5.282586516 podStartE2EDuration="7.729115572s" podCreationTimestamp="2025-12-08 19:36:58 +0000 UTC" firstStartedPulling="2025-12-08 19:37:02.673287309 +0000 UTC m=+905.315488312" lastFinishedPulling="2025-12-08 19:37:05.119816365 +0000 UTC m=+907.762017368" observedRunningTime="2025-12-08 19:37:05.727425175 +0000 UTC m=+908.369626188" watchObservedRunningTime="2025-12-08 19:37:05.729115572 +0000 UTC m=+908.371316575" Dec 08 19:37:06 crc kubenswrapper[4706]: I1208 19:37:06.720459 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jtzsv" event={"ID":"1af879c2-1138-4ca6-92d5-013c50cb4ef2","Type":"ContainerStarted","Data":"0d89edd1c7f5f8fb04581c491f52e12416867bc973103da0af2054ffa4d6ee05"} Dec 08 19:37:06 crc kubenswrapper[4706]: I1208 19:37:06.752225 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jtzsv" podStartSLOduration=5.561591507 podStartE2EDuration="14.752204581s" podCreationTimestamp="2025-12-08 19:36:52 +0000 UTC" firstStartedPulling="2025-12-08 19:36:52.821693246 +0000 UTC m=+895.463894249" lastFinishedPulling="2025-12-08 19:37:02.01230632 +0000 UTC m=+904.654507323" observedRunningTime="2025-12-08 19:37:06.751089029 +0000 UTC m=+909.393290032" watchObservedRunningTime="2025-12-08 19:37:06.752204581 +0000 UTC m=+909.394405584" Dec 08 19:37:07 crc kubenswrapper[4706]: I1208 19:37:07.678639 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:37:07 crc kubenswrapper[4706]: I1208 19:37:07.718827 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:37:07 crc kubenswrapper[4706]: I1208 19:37:07.728057 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:37:08 crc kubenswrapper[4706]: I1208 19:37:08.814592 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:37:08 crc kubenswrapper[4706]: I1208 19:37:08.815073 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:37:08 crc kubenswrapper[4706]: I1208 19:37:08.888814 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:37:12 crc kubenswrapper[4706]: I1208 19:37:12.392418 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mdfmz" Dec 08 19:37:13 crc kubenswrapper[4706]: I1208 19:37:13.417427 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:37:13 crc kubenswrapper[4706]: I1208 19:37:13.476118 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:37:13 crc kubenswrapper[4706]: I1208 19:37:13.993756 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-49kgz" Dec 08 19:37:15 crc kubenswrapper[4706]: I1208 19:37:15.836475 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtffp"] Dec 08 19:37:15 crc kubenswrapper[4706]: I1208 19:37:15.837168 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dtffp" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="registry-server" containerID="cri-o://5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608" gracePeriod=2 Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.277274 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.347974 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98dv6\" (UniqueName: \"kubernetes.io/projected/c8a77b1e-cb15-4263-91f2-a437dafb360e-kube-api-access-98dv6\") pod \"c8a77b1e-cb15-4263-91f2-a437dafb360e\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.348084 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-catalog-content\") pod \"c8a77b1e-cb15-4263-91f2-a437dafb360e\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.348134 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-utilities\") pod \"c8a77b1e-cb15-4263-91f2-a437dafb360e\" (UID: \"c8a77b1e-cb15-4263-91f2-a437dafb360e\") " Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.349384 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-utilities" (OuterVolumeSpecName: "utilities") pod "c8a77b1e-cb15-4263-91f2-a437dafb360e" (UID: "c8a77b1e-cb15-4263-91f2-a437dafb360e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.355795 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8a77b1e-cb15-4263-91f2-a437dafb360e-kube-api-access-98dv6" (OuterVolumeSpecName: "kube-api-access-98dv6") pod "c8a77b1e-cb15-4263-91f2-a437dafb360e" (UID: "c8a77b1e-cb15-4263-91f2-a437dafb360e"). InnerVolumeSpecName "kube-api-access-98dv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.367530 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8a77b1e-cb15-4263-91f2-a437dafb360e" (UID: "c8a77b1e-cb15-4263-91f2-a437dafb360e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.451293 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98dv6\" (UniqueName: \"kubernetes.io/projected/c8a77b1e-cb15-4263-91f2-a437dafb360e-kube-api-access-98dv6\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.451338 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.451348 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a77b1e-cb15-4263-91f2-a437dafb360e-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.791582 4706 generic.go:334] "Generic (PLEG): container finished" podID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerID="5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608" exitCode=0 Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.791644 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtffp" event={"ID":"c8a77b1e-cb15-4263-91f2-a437dafb360e","Type":"ContainerDied","Data":"5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608"} Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.791689 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtffp" event={"ID":"c8a77b1e-cb15-4263-91f2-a437dafb360e","Type":"ContainerDied","Data":"73a27ad033cb13de11b8f0cde6856d0aedec9db5398e1e0ef26c67bd85b732a2"} Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.791682 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtffp" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.791712 4706 scope.go:117] "RemoveContainer" containerID="5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.814643 4706 scope.go:117] "RemoveContainer" containerID="37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.827097 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtffp"] Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.833714 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtffp"] Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.843669 4706 scope.go:117] "RemoveContainer" containerID="f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.861651 4706 scope.go:117] "RemoveContainer" containerID="5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608" Dec 08 19:37:16 crc kubenswrapper[4706]: E1208 19:37:16.862299 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608\": container with ID starting with 5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608 not found: ID does not exist" containerID="5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.862368 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608"} err="failed to get container status \"5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608\": rpc error: code = NotFound desc = could not find container \"5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608\": container with ID starting with 5c0735e9ec7830ae9d1cd7190121c11cb81b1fd4504d8fac2344394f1e339608 not found: ID does not exist" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.862403 4706 scope.go:117] "RemoveContainer" containerID="37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa" Dec 08 19:37:16 crc kubenswrapper[4706]: E1208 19:37:16.862796 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa\": container with ID starting with 37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa not found: ID does not exist" containerID="37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.862845 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa"} err="failed to get container status \"37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa\": rpc error: code = NotFound desc = could not find container \"37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa\": container with ID starting with 37e323aa166ca982cdf14d5abb3c8e1f6fb9fe8ba86cfa1ed79b36b284bcf0fa not found: ID does not exist" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.862881 4706 scope.go:117] "RemoveContainer" containerID="f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf" Dec 08 19:37:16 crc kubenswrapper[4706]: E1208 19:37:16.863212 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf\": container with ID starting with f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf not found: ID does not exist" containerID="f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf" Dec 08 19:37:16 crc kubenswrapper[4706]: I1208 19:37:16.863291 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf"} err="failed to get container status \"f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf\": rpc error: code = NotFound desc = could not find container \"f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf\": container with ID starting with f06124b8e0246e7650c3b168bb9baed2be7831595053a125bb303932c9cecbcf not found: ID does not exist" Dec 08 19:37:17 crc kubenswrapper[4706]: I1208 19:37:17.640174 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" path="/var/lib/kubelet/pods/c8a77b1e-cb15-4263-91f2-a437dafb360e/volumes" Dec 08 19:37:18 crc kubenswrapper[4706]: I1208 19:37:18.875948 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.251607 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-c5fr6"] Dec 08 19:37:21 crc kubenswrapper[4706]: E1208 19:37:21.251914 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="extract-content" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.251929 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="extract-content" Dec 08 19:37:21 crc kubenswrapper[4706]: E1208 19:37:21.251948 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="extract-utilities" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.251957 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="extract-utilities" Dec 08 19:37:21 crc kubenswrapper[4706]: E1208 19:37:21.251965 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="registry-server" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.251971 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="registry-server" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.252083 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8a77b1e-cb15-4263-91f2-a437dafb360e" containerName="registry-server" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.252555 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.254759 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.254839 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.255064 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-9qq48" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.264287 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-c5fr6"] Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.329304 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mprz2\" (UniqueName: \"kubernetes.io/projected/ca6c738d-aa2e-4be3-9827-fc03af84458d-kube-api-access-mprz2\") pod \"openstack-operator-index-c5fr6\" (UID: \"ca6c738d-aa2e-4be3-9827-fc03af84458d\") " pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.431415 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mprz2\" (UniqueName: \"kubernetes.io/projected/ca6c738d-aa2e-4be3-9827-fc03af84458d-kube-api-access-mprz2\") pod \"openstack-operator-index-c5fr6\" (UID: \"ca6c738d-aa2e-4be3-9827-fc03af84458d\") " pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.451850 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mprz2\" (UniqueName: \"kubernetes.io/projected/ca6c738d-aa2e-4be3-9827-fc03af84458d-kube-api-access-mprz2\") pod \"openstack-operator-index-c5fr6\" (UID: \"ca6c738d-aa2e-4be3-9827-fc03af84458d\") " pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:21 crc kubenswrapper[4706]: I1208 19:37:21.577188 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:22 crc kubenswrapper[4706]: I1208 19:37:22.053594 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-c5fr6"] Dec 08 19:37:22 crc kubenswrapper[4706]: W1208 19:37:22.065660 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca6c738d_aa2e_4be3_9827_fc03af84458d.slice/crio-7a1f3f9c0ab2eab1e40523134d39f63f1a5a34b49b2e08e8bfe5f06924a49403 WatchSource:0}: Error finding container 7a1f3f9c0ab2eab1e40523134d39f63f1a5a34b49b2e08e8bfe5f06924a49403: Status 404 returned error can't find the container with id 7a1f3f9c0ab2eab1e40523134d39f63f1a5a34b49b2e08e8bfe5f06924a49403 Dec 08 19:37:22 crc kubenswrapper[4706]: I1208 19:37:22.680812 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jtzsv" Dec 08 19:37:22 crc kubenswrapper[4706]: I1208 19:37:22.841111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-c5fr6" event={"ID":"ca6c738d-aa2e-4be3-9827-fc03af84458d","Type":"ContainerStarted","Data":"7a1f3f9c0ab2eab1e40523134d39f63f1a5a34b49b2e08e8bfe5f06924a49403"} Dec 08 19:37:24 crc kubenswrapper[4706]: I1208 19:37:24.856565 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-c5fr6" event={"ID":"ca6c738d-aa2e-4be3-9827-fc03af84458d","Type":"ContainerStarted","Data":"b6e4ce107a78aa3cd201f1fca3ae1cb8339f46e972003c426cb3c2a4b84cc9a6"} Dec 08 19:37:24 crc kubenswrapper[4706]: I1208 19:37:24.874996 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-c5fr6" podStartSLOduration=1.35641629 podStartE2EDuration="3.874970036s" podCreationTimestamp="2025-12-08 19:37:21 +0000 UTC" firstStartedPulling="2025-12-08 19:37:22.068162309 +0000 UTC m=+924.710363312" lastFinishedPulling="2025-12-08 19:37:24.586716055 +0000 UTC m=+927.228917058" observedRunningTime="2025-12-08 19:37:24.869697597 +0000 UTC m=+927.511898600" watchObservedRunningTime="2025-12-08 19:37:24.874970036 +0000 UTC m=+927.517171039" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.075195 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8776b"] Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.077011 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.098739 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-utilities\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.098810 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8bq7\" (UniqueName: \"kubernetes.io/projected/40b7e145-0275-459f-babb-caff5c9802cd-kube-api-access-d8bq7\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.098861 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-catalog-content\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.103717 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8776b"] Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.200813 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-utilities\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.200884 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8bq7\" (UniqueName: \"kubernetes.io/projected/40b7e145-0275-459f-babb-caff5c9802cd-kube-api-access-d8bq7\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.200935 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-catalog-content\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.201460 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-utilities\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.201551 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-catalog-content\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.242526 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8bq7\" (UniqueName: \"kubernetes.io/projected/40b7e145-0275-459f-babb-caff5c9802cd-kube-api-access-d8bq7\") pod \"community-operators-8776b\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.411343 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:25 crc kubenswrapper[4706]: I1208 19:37:25.944438 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8776b"] Dec 08 19:37:26 crc kubenswrapper[4706]: I1208 19:37:26.875644 4706 generic.go:334] "Generic (PLEG): container finished" podID="40b7e145-0275-459f-babb-caff5c9802cd" containerID="9da102eddedfc5f349c5d0c4d152a9f611087a1520572d354ef4a144bc59e25a" exitCode=0 Dec 08 19:37:26 crc kubenswrapper[4706]: I1208 19:37:26.875730 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8776b" event={"ID":"40b7e145-0275-459f-babb-caff5c9802cd","Type":"ContainerDied","Data":"9da102eddedfc5f349c5d0c4d152a9f611087a1520572d354ef4a144bc59e25a"} Dec 08 19:37:26 crc kubenswrapper[4706]: I1208 19:37:26.876055 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8776b" event={"ID":"40b7e145-0275-459f-babb-caff5c9802cd","Type":"ContainerStarted","Data":"f0282f9de74983ab9c62764dbde369293fae91e836e7aa427a83e492ec7aff39"} Dec 08 19:37:27 crc kubenswrapper[4706]: I1208 19:37:27.837459 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bsh55"] Dec 08 19:37:27 crc kubenswrapper[4706]: I1208 19:37:27.838194 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bsh55" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerName="registry-server" containerID="cri-o://c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b" gracePeriod=2 Dec 08 19:37:27 crc kubenswrapper[4706]: I1208 19:37:27.886310 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8776b" event={"ID":"40b7e145-0275-459f-babb-caff5c9802cd","Type":"ContainerStarted","Data":"72df8d15c56a4e99047ab67fbcf3d58f2167583cbf52f3b5f34fe992e13931d2"} Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.242544 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.246818 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-catalog-content\") pod \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.246958 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gkwr\" (UniqueName: \"kubernetes.io/projected/a6913af1-1b8d-4d14-8b41-7f44443f88f6-kube-api-access-5gkwr\") pod \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.247119 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-utilities\") pod \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\" (UID: \"a6913af1-1b8d-4d14-8b41-7f44443f88f6\") " Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.248097 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-utilities" (OuterVolumeSpecName: "utilities") pod "a6913af1-1b8d-4d14-8b41-7f44443f88f6" (UID: "a6913af1-1b8d-4d14-8b41-7f44443f88f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.254373 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6913af1-1b8d-4d14-8b41-7f44443f88f6-kube-api-access-5gkwr" (OuterVolumeSpecName: "kube-api-access-5gkwr") pod "a6913af1-1b8d-4d14-8b41-7f44443f88f6" (UID: "a6913af1-1b8d-4d14-8b41-7f44443f88f6"). InnerVolumeSpecName "kube-api-access-5gkwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.305538 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6913af1-1b8d-4d14-8b41-7f44443f88f6" (UID: "a6913af1-1b8d-4d14-8b41-7f44443f88f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.348692 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.348729 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6913af1-1b8d-4d14-8b41-7f44443f88f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.348747 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gkwr\" (UniqueName: \"kubernetes.io/projected/a6913af1-1b8d-4d14-8b41-7f44443f88f6-kube-api-access-5gkwr\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.896159 4706 generic.go:334] "Generic (PLEG): container finished" podID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerID="c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b" exitCode=0 Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.896234 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bsh55" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.896246 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsh55" event={"ID":"a6913af1-1b8d-4d14-8b41-7f44443f88f6","Type":"ContainerDied","Data":"c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b"} Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.896346 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsh55" event={"ID":"a6913af1-1b8d-4d14-8b41-7f44443f88f6","Type":"ContainerDied","Data":"45623180e411d2b9c08346386e0b98e0a599816f86438f9e409f9ee0690da54b"} Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.896373 4706 scope.go:117] "RemoveContainer" containerID="c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.899091 4706 generic.go:334] "Generic (PLEG): container finished" podID="40b7e145-0275-459f-babb-caff5c9802cd" containerID="72df8d15c56a4e99047ab67fbcf3d58f2167583cbf52f3b5f34fe992e13931d2" exitCode=0 Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.899125 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8776b" event={"ID":"40b7e145-0275-459f-babb-caff5c9802cd","Type":"ContainerDied","Data":"72df8d15c56a4e99047ab67fbcf3d58f2167583cbf52f3b5f34fe992e13931d2"} Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.920741 4706 scope.go:117] "RemoveContainer" containerID="df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.946924 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bsh55"] Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.953413 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bsh55"] Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.955066 4706 scope.go:117] "RemoveContainer" containerID="f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.974592 4706 scope.go:117] "RemoveContainer" containerID="c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b" Dec 08 19:37:28 crc kubenswrapper[4706]: E1208 19:37:28.975455 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b\": container with ID starting with c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b not found: ID does not exist" containerID="c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.975529 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b"} err="failed to get container status \"c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b\": rpc error: code = NotFound desc = could not find container \"c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b\": container with ID starting with c6457266a8e1b7e801605106048fef741202b900ccb5ce5e5ff4eb7d33d8cb3b not found: ID does not exist" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.975575 4706 scope.go:117] "RemoveContainer" containerID="df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6" Dec 08 19:37:28 crc kubenswrapper[4706]: E1208 19:37:28.976137 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6\": container with ID starting with df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6 not found: ID does not exist" containerID="df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.976179 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6"} err="failed to get container status \"df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6\": rpc error: code = NotFound desc = could not find container \"df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6\": container with ID starting with df9698f7502d84ccaff26f8c3a02594293e1dad59bf3c7e580215fef444bedf6 not found: ID does not exist" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.976206 4706 scope.go:117] "RemoveContainer" containerID="f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842" Dec 08 19:37:28 crc kubenswrapper[4706]: E1208 19:37:28.976708 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842\": container with ID starting with f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842 not found: ID does not exist" containerID="f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842" Dec 08 19:37:28 crc kubenswrapper[4706]: I1208 19:37:28.976755 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842"} err="failed to get container status \"f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842\": rpc error: code = NotFound desc = could not find container \"f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842\": container with ID starting with f908c110020ec4a53d4ebc27e3afe8a9d44ea4649671b2fe8b570fd74eadb842 not found: ID does not exist" Dec 08 19:37:29 crc kubenswrapper[4706]: I1208 19:37:29.662573 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" path="/var/lib/kubelet/pods/a6913af1-1b8d-4d14-8b41-7f44443f88f6/volumes" Dec 08 19:37:29 crc kubenswrapper[4706]: I1208 19:37:29.910885 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8776b" event={"ID":"40b7e145-0275-459f-babb-caff5c9802cd","Type":"ContainerStarted","Data":"a4b86d85e243e29b2fc53f77f0bb32a4cc3ee80afc571cc6facb7fb77b4ca54c"} Dec 08 19:37:29 crc kubenswrapper[4706]: I1208 19:37:29.943477 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8776b" podStartSLOduration=2.469329057 podStartE2EDuration="4.943453516s" podCreationTimestamp="2025-12-08 19:37:25 +0000 UTC" firstStartedPulling="2025-12-08 19:37:26.877696737 +0000 UTC m=+929.519897740" lastFinishedPulling="2025-12-08 19:37:29.351821206 +0000 UTC m=+931.994022199" observedRunningTime="2025-12-08 19:37:29.940012129 +0000 UTC m=+932.582213132" watchObservedRunningTime="2025-12-08 19:37:29.943453516 +0000 UTC m=+932.585654519" Dec 08 19:37:31 crc kubenswrapper[4706]: I1208 19:37:31.578393 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:31 crc kubenswrapper[4706]: I1208 19:37:31.578882 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:31 crc kubenswrapper[4706]: I1208 19:37:31.616461 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:31 crc kubenswrapper[4706]: I1208 19:37:31.950369 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-c5fr6" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.907773 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8"] Dec 08 19:37:34 crc kubenswrapper[4706]: E1208 19:37:34.908554 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerName="extract-content" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.908576 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerName="extract-content" Dec 08 19:37:34 crc kubenswrapper[4706]: E1208 19:37:34.908591 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerName="registry-server" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.908599 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerName="registry-server" Dec 08 19:37:34 crc kubenswrapper[4706]: E1208 19:37:34.908625 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerName="extract-utilities" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.908633 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerName="extract-utilities" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.908780 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6913af1-1b8d-4d14-8b41-7f44443f88f6" containerName="registry-server" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.909775 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.914919 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-4wbbz" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.929348 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8"] Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.955054 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-util\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.955323 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6qhs\" (UniqueName: \"kubernetes.io/projected/3231e588-6c34-4c26-9909-ece08e2661c3-kube-api-access-k6qhs\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:34 crc kubenswrapper[4706]: I1208 19:37:34.955388 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-bundle\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.056151 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6qhs\" (UniqueName: \"kubernetes.io/projected/3231e588-6c34-4c26-9909-ece08e2661c3-kube-api-access-k6qhs\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.056245 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-bundle\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.056335 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-util\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.056782 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-util\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.056811 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-bundle\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.082650 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6qhs\" (UniqueName: \"kubernetes.io/projected/3231e588-6c34-4c26-9909-ece08e2661c3-kube-api-access-k6qhs\") pod \"dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.227627 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.412659 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.415389 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.470345 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.503800 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8"] Dec 08 19:37:35 crc kubenswrapper[4706]: I1208 19:37:35.960776 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" event={"ID":"3231e588-6c34-4c26-9909-ece08e2661c3","Type":"ContainerStarted","Data":"ad4aa166320cdf501c3e2ba6abcba2791f5a3cc10b8b24d0c1f8373c30d19a05"} Dec 08 19:37:36 crc kubenswrapper[4706]: I1208 19:37:36.006785 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:36 crc kubenswrapper[4706]: I1208 19:37:36.977075 4706 generic.go:334] "Generic (PLEG): container finished" podID="3231e588-6c34-4c26-9909-ece08e2661c3" containerID="f93a74ac5319019960f5f427e4eb3f0f7483c9862ec10c1dab0bcc0abc4a29bb" exitCode=0 Dec 08 19:37:36 crc kubenswrapper[4706]: I1208 19:37:36.977146 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" event={"ID":"3231e588-6c34-4c26-9909-ece08e2661c3","Type":"ContainerDied","Data":"f93a74ac5319019960f5f427e4eb3f0f7483c9862ec10c1dab0bcc0abc4a29bb"} Dec 08 19:37:36 crc kubenswrapper[4706]: I1208 19:37:36.981513 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 19:37:37 crc kubenswrapper[4706]: I1208 19:37:37.038305 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8776b"] Dec 08 19:37:37 crc kubenswrapper[4706]: I1208 19:37:37.987937 4706 generic.go:334] "Generic (PLEG): container finished" podID="3231e588-6c34-4c26-9909-ece08e2661c3" containerID="1f4ddf72b1dc730b8c495fe4c9d5699b66b8f33119ca7d09acfe741313fe5bad" exitCode=0 Dec 08 19:37:37 crc kubenswrapper[4706]: I1208 19:37:37.987999 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" event={"ID":"3231e588-6c34-4c26-9909-ece08e2661c3","Type":"ContainerDied","Data":"1f4ddf72b1dc730b8c495fe4c9d5699b66b8f33119ca7d09acfe741313fe5bad"} Dec 08 19:37:39 crc kubenswrapper[4706]: I1208 19:37:39.001199 4706 generic.go:334] "Generic (PLEG): container finished" podID="3231e588-6c34-4c26-9909-ece08e2661c3" containerID="f2f810fb95b6dab0418e5080bd6f7ea75d8fae065ef411472be7c1aba5b0f6a2" exitCode=0 Dec 08 19:37:39 crc kubenswrapper[4706]: I1208 19:37:39.001288 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" event={"ID":"3231e588-6c34-4c26-9909-ece08e2661c3","Type":"ContainerDied","Data":"f2f810fb95b6dab0418e5080bd6f7ea75d8fae065ef411472be7c1aba5b0f6a2"} Dec 08 19:37:39 crc kubenswrapper[4706]: I1208 19:37:39.001642 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8776b" podUID="40b7e145-0275-459f-babb-caff5c9802cd" containerName="registry-server" containerID="cri-o://a4b86d85e243e29b2fc53f77f0bb32a4cc3ee80afc571cc6facb7fb77b4ca54c" gracePeriod=2 Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.009792 4706 generic.go:334] "Generic (PLEG): container finished" podID="40b7e145-0275-459f-babb-caff5c9802cd" containerID="a4b86d85e243e29b2fc53f77f0bb32a4cc3ee80afc571cc6facb7fb77b4ca54c" exitCode=0 Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.009893 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8776b" event={"ID":"40b7e145-0275-459f-babb-caff5c9802cd","Type":"ContainerDied","Data":"a4b86d85e243e29b2fc53f77f0bb32a4cc3ee80afc571cc6facb7fb77b4ca54c"} Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.010289 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8776b" event={"ID":"40b7e145-0275-459f-babb-caff5c9802cd","Type":"ContainerDied","Data":"f0282f9de74983ab9c62764dbde369293fae91e836e7aa427a83e492ec7aff39"} Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.010322 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0282f9de74983ab9c62764dbde369293fae91e836e7aa427a83e492ec7aff39" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.042379 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.154918 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-utilities\") pod \"40b7e145-0275-459f-babb-caff5c9802cd\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.154995 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-catalog-content\") pod \"40b7e145-0275-459f-babb-caff5c9802cd\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.155091 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8bq7\" (UniqueName: \"kubernetes.io/projected/40b7e145-0275-459f-babb-caff5c9802cd-kube-api-access-d8bq7\") pod \"40b7e145-0275-459f-babb-caff5c9802cd\" (UID: \"40b7e145-0275-459f-babb-caff5c9802cd\") " Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.156089 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-utilities" (OuterVolumeSpecName: "utilities") pod "40b7e145-0275-459f-babb-caff5c9802cd" (UID: "40b7e145-0275-459f-babb-caff5c9802cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.165425 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40b7e145-0275-459f-babb-caff5c9802cd-kube-api-access-d8bq7" (OuterVolumeSpecName: "kube-api-access-d8bq7") pod "40b7e145-0275-459f-babb-caff5c9802cd" (UID: "40b7e145-0275-459f-babb-caff5c9802cd"). InnerVolumeSpecName "kube-api-access-d8bq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.211769 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40b7e145-0275-459f-babb-caff5c9802cd" (UID: "40b7e145-0275-459f-babb-caff5c9802cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.256400 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.256444 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b7e145-0275-459f-babb-caff5c9802cd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.256461 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8bq7\" (UniqueName: \"kubernetes.io/projected/40b7e145-0275-459f-babb-caff5c9802cd-kube-api-access-d8bq7\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.340299 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.459439 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6qhs\" (UniqueName: \"kubernetes.io/projected/3231e588-6c34-4c26-9909-ece08e2661c3-kube-api-access-k6qhs\") pod \"3231e588-6c34-4c26-9909-ece08e2661c3\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.459950 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-util\") pod \"3231e588-6c34-4c26-9909-ece08e2661c3\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.460070 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-bundle\") pod \"3231e588-6c34-4c26-9909-ece08e2661c3\" (UID: \"3231e588-6c34-4c26-9909-ece08e2661c3\") " Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.461096 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-bundle" (OuterVolumeSpecName: "bundle") pod "3231e588-6c34-4c26-9909-ece08e2661c3" (UID: "3231e588-6c34-4c26-9909-ece08e2661c3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.464377 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3231e588-6c34-4c26-9909-ece08e2661c3-kube-api-access-k6qhs" (OuterVolumeSpecName: "kube-api-access-k6qhs") pod "3231e588-6c34-4c26-9909-ece08e2661c3" (UID: "3231e588-6c34-4c26-9909-ece08e2661c3"). InnerVolumeSpecName "kube-api-access-k6qhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.477531 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-util" (OuterVolumeSpecName: "util") pod "3231e588-6c34-4c26-9909-ece08e2661c3" (UID: "3231e588-6c34-4c26-9909-ece08e2661c3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.561510 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-util\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.561557 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3231e588-6c34-4c26-9909-ece08e2661c3-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:40 crc kubenswrapper[4706]: I1208 19:37:40.561570 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6qhs\" (UniqueName: \"kubernetes.io/projected/3231e588-6c34-4c26-9909-ece08e2661c3-kube-api-access-k6qhs\") on node \"crc\" DevicePath \"\"" Dec 08 19:37:41 crc kubenswrapper[4706]: I1208 19:37:41.019695 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" event={"ID":"3231e588-6c34-4c26-9909-ece08e2661c3","Type":"ContainerDied","Data":"ad4aa166320cdf501c3e2ba6abcba2791f5a3cc10b8b24d0c1f8373c30d19a05"} Dec 08 19:37:41 crc kubenswrapper[4706]: I1208 19:37:41.019744 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8776b" Dec 08 19:37:41 crc kubenswrapper[4706]: I1208 19:37:41.019761 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad4aa166320cdf501c3e2ba6abcba2791f5a3cc10b8b24d0c1f8373c30d19a05" Dec 08 19:37:41 crc kubenswrapper[4706]: I1208 19:37:41.019803 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8" Dec 08 19:37:41 crc kubenswrapper[4706]: I1208 19:37:41.072390 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8776b"] Dec 08 19:37:41 crc kubenswrapper[4706]: I1208 19:37:41.085400 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8776b"] Dec 08 19:37:41 crc kubenswrapper[4706]: I1208 19:37:41.618003 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40b7e145-0275-459f-babb-caff5c9802cd" path="/var/lib/kubelet/pods/40b7e145-0275-459f-babb-caff5c9802cd/volumes" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.665577 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p"] Dec 08 19:37:45 crc kubenswrapper[4706]: E1208 19:37:45.666707 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b7e145-0275-459f-babb-caff5c9802cd" containerName="extract-utilities" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.666759 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b7e145-0275-459f-babb-caff5c9802cd" containerName="extract-utilities" Dec 08 19:37:45 crc kubenswrapper[4706]: E1208 19:37:45.666773 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3231e588-6c34-4c26-9909-ece08e2661c3" containerName="extract" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.666779 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3231e588-6c34-4c26-9909-ece08e2661c3" containerName="extract" Dec 08 19:37:45 crc kubenswrapper[4706]: E1208 19:37:45.666792 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b7e145-0275-459f-babb-caff5c9802cd" containerName="registry-server" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.666798 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b7e145-0275-459f-babb-caff5c9802cd" containerName="registry-server" Dec 08 19:37:45 crc kubenswrapper[4706]: E1208 19:37:45.666806 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b7e145-0275-459f-babb-caff5c9802cd" containerName="extract-content" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.666813 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b7e145-0275-459f-babb-caff5c9802cd" containerName="extract-content" Dec 08 19:37:45 crc kubenswrapper[4706]: E1208 19:37:45.666828 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3231e588-6c34-4c26-9909-ece08e2661c3" containerName="pull" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.666834 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3231e588-6c34-4c26-9909-ece08e2661c3" containerName="pull" Dec 08 19:37:45 crc kubenswrapper[4706]: E1208 19:37:45.666848 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3231e588-6c34-4c26-9909-ece08e2661c3" containerName="util" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.666854 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3231e588-6c34-4c26-9909-ece08e2661c3" containerName="util" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.666976 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3231e588-6c34-4c26-9909-ece08e2661c3" containerName="extract" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.666995 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b7e145-0275-459f-babb-caff5c9802cd" containerName="registry-server" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.667562 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.671632 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-dtz9k" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.701007 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p"] Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.851169 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldpl\" (UniqueName: \"kubernetes.io/projected/ea53cbe8-559b-45be-8ae8-25aeac554de4-kube-api-access-bldpl\") pod \"openstack-operator-controller-operator-6d96469866-85q7p\" (UID: \"ea53cbe8-559b-45be-8ae8-25aeac554de4\") " pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.952747 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bldpl\" (UniqueName: \"kubernetes.io/projected/ea53cbe8-559b-45be-8ae8-25aeac554de4-kube-api-access-bldpl\") pod \"openstack-operator-controller-operator-6d96469866-85q7p\" (UID: \"ea53cbe8-559b-45be-8ae8-25aeac554de4\") " pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.974535 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldpl\" (UniqueName: \"kubernetes.io/projected/ea53cbe8-559b-45be-8ae8-25aeac554de4-kube-api-access-bldpl\") pod \"openstack-operator-controller-operator-6d96469866-85q7p\" (UID: \"ea53cbe8-559b-45be-8ae8-25aeac554de4\") " pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" Dec 08 19:37:45 crc kubenswrapper[4706]: I1208 19:37:45.995620 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" Dec 08 19:37:46 crc kubenswrapper[4706]: I1208 19:37:46.454829 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p"] Dec 08 19:37:47 crc kubenswrapper[4706]: I1208 19:37:47.075901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" event={"ID":"ea53cbe8-559b-45be-8ae8-25aeac554de4","Type":"ContainerStarted","Data":"560300d82dd94338029a3b66c83f2a9d89f8bfb8b84429bbf958bf89912b9702"} Dec 08 19:37:51 crc kubenswrapper[4706]: I1208 19:37:51.109508 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" event={"ID":"ea53cbe8-559b-45be-8ae8-25aeac554de4","Type":"ContainerStarted","Data":"95b42f1b6d3ebd8a1f9cb1d56fd0cfa48720fcd531a56ecf0cd6df29527c0804"} Dec 08 19:37:51 crc kubenswrapper[4706]: I1208 19:37:51.110288 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" Dec 08 19:37:51 crc kubenswrapper[4706]: I1208 19:37:51.141655 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" podStartSLOduration=2.060455809 podStartE2EDuration="6.141636477s" podCreationTimestamp="2025-12-08 19:37:45 +0000 UTC" firstStartedPulling="2025-12-08 19:37:46.471730271 +0000 UTC m=+949.113931274" lastFinishedPulling="2025-12-08 19:37:50.552910939 +0000 UTC m=+953.195111942" observedRunningTime="2025-12-08 19:37:51.139023363 +0000 UTC m=+953.781224366" watchObservedRunningTime="2025-12-08 19:37:51.141636477 +0000 UTC m=+953.783837480" Dec 08 19:37:55 crc kubenswrapper[4706]: I1208 19:37:55.998438 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6d96469866-85q7p" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.860649 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw"] Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.864148 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.867054 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-9zcjj" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.870750 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl"] Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.872814 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.876410 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-gz4k6" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.882509 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw"] Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.896442 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs"] Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.897691 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.900167 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-sxmbg" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.917644 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jcvs\" (UniqueName: \"kubernetes.io/projected/4c42f8b9-a85c-4a45-a8f8-4a43e561a14e-kube-api-access-4jcvs\") pod \"barbican-operator-controller-manager-7d9dfd778-qtnlw\" (UID: \"4c42f8b9-a85c-4a45-a8f8-4a43e561a14e\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.917728 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssbxj\" (UniqueName: \"kubernetes.io/projected/0f144ce9-7cc0-41f3-84ea-654579d32c0b-kube-api-access-ssbxj\") pod \"designate-operator-controller-manager-697fb699cf-7fvjs\" (UID: \"0f144ce9-7cc0-41f3-84ea-654579d32c0b\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.917872 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzjxw\" (UniqueName: \"kubernetes.io/projected/e5899e67-6618-43f5-8454-fddfae494e4a-kube-api-access-gzjxw\") pod \"cinder-operator-controller-manager-6c677c69b-j2mjl\" (UID: \"e5899e67-6618-43f5-8454-fddfae494e4a\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.939242 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs"] Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.943912 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl"] Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.964402 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l"] Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.965738 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.969374 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-brqbb" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.978365 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l"] Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.979988 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.995940 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4vcrx" Dec 08 19:38:14 crc kubenswrapper[4706]: I1208 19:38:14.996174 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.002187 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.014366 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.015914 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.021339 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.022855 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjhfr\" (UniqueName: \"kubernetes.io/projected/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-kube-api-access-rjhfr\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.022924 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flrdg\" (UniqueName: \"kubernetes.io/projected/2acd6f54-87a9-4e50-859c-42a7c8e9d098-kube-api-access-flrdg\") pod \"heat-operator-controller-manager-5f64f6f8bb-cmg9l\" (UID: \"2acd6f54-87a9-4e50-859c-42a7c8e9d098\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.022971 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.023019 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzjxw\" (UniqueName: \"kubernetes.io/projected/e5899e67-6618-43f5-8454-fddfae494e4a-kube-api-access-gzjxw\") pod \"cinder-operator-controller-manager-6c677c69b-j2mjl\" (UID: \"e5899e67-6618-43f5-8454-fddfae494e4a\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.023025 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.023054 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqtzs\" (UniqueName: \"kubernetes.io/projected/b02b307a-7d4c-4d20-b388-18f6fc6ec0ee-kube-api-access-lqtzs\") pod \"glance-operator-controller-manager-5697bb5779-b7m9l\" (UID: \"b02b307a-7d4c-4d20-b388-18f6fc6ec0ee\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.023172 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jcvs\" (UniqueName: \"kubernetes.io/projected/4c42f8b9-a85c-4a45-a8f8-4a43e561a14e-kube-api-access-4jcvs\") pod \"barbican-operator-controller-manager-7d9dfd778-qtnlw\" (UID: \"4c42f8b9-a85c-4a45-a8f8-4a43e561a14e\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.023207 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssbxj\" (UniqueName: \"kubernetes.io/projected/0f144ce9-7cc0-41f3-84ea-654579d32c0b-kube-api-access-ssbxj\") pod \"designate-operator-controller-manager-697fb699cf-7fvjs\" (UID: \"0f144ce9-7cc0-41f3-84ea-654579d32c0b\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.024585 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.024947 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-6kp2h" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.028721 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-fktxr" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.030780 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.051345 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.068566 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssbxj\" (UniqueName: \"kubernetes.io/projected/0f144ce9-7cc0-41f3-84ea-654579d32c0b-kube-api-access-ssbxj\") pod \"designate-operator-controller-manager-697fb699cf-7fvjs\" (UID: \"0f144ce9-7cc0-41f3-84ea-654579d32c0b\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.069771 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzjxw\" (UniqueName: \"kubernetes.io/projected/e5899e67-6618-43f5-8454-fddfae494e4a-kube-api-access-gzjxw\") pod \"cinder-operator-controller-manager-6c677c69b-j2mjl\" (UID: \"e5899e67-6618-43f5-8454-fddfae494e4a\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.078455 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jcvs\" (UniqueName: \"kubernetes.io/projected/4c42f8b9-a85c-4a45-a8f8-4a43e561a14e-kube-api-access-4jcvs\") pod \"barbican-operator-controller-manager-7d9dfd778-qtnlw\" (UID: \"4c42f8b9-a85c-4a45-a8f8-4a43e561a14e\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.118372 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.120121 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.129711 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-g6n74" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.133468 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjhfr\" (UniqueName: \"kubernetes.io/projected/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-kube-api-access-rjhfr\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.133521 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flrdg\" (UniqueName: \"kubernetes.io/projected/2acd6f54-87a9-4e50-859c-42a7c8e9d098-kube-api-access-flrdg\") pod \"heat-operator-controller-manager-5f64f6f8bb-cmg9l\" (UID: \"2acd6f54-87a9-4e50-859c-42a7c8e9d098\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.133574 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.133636 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqtzs\" (UniqueName: \"kubernetes.io/projected/b02b307a-7d4c-4d20-b388-18f6fc6ec0ee-kube-api-access-lqtzs\") pod \"glance-operator-controller-manager-5697bb5779-b7m9l\" (UID: \"b02b307a-7d4c-4d20-b388-18f6fc6ec0ee\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.133684 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2csgw\" (UniqueName: \"kubernetes.io/projected/fab26e20-8c1c-4a9e-b7c4-8d4815a1568e-kube-api-access-2csgw\") pod \"keystone-operator-controller-manager-7765d96ddf-kg2wq\" (UID: \"fab26e20-8c1c-4a9e-b7c4-8d4815a1568e\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.133724 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66shv\" (UniqueName: \"kubernetes.io/projected/1290df78-3e2c-494e-a264-c495d80c0bef-kube-api-access-66shv\") pod \"horizon-operator-controller-manager-68c6d99b8f-cggng\" (UID: \"1290df78-3e2c-494e-a264-c495d80c0bef\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" Dec 08 19:38:15 crc kubenswrapper[4706]: E1208 19:38:15.134414 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:15 crc kubenswrapper[4706]: E1208 19:38:15.134468 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert podName:ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:15.634450861 +0000 UTC m=+978.276651864 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert") pod "infra-operator-controller-manager-78d48bff9d-qmcdh" (UID: "ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3") : secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.137148 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.142197 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.157366 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.158762 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-dzqf8" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.165579 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.166833 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.181051 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-g68qq" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.193225 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flrdg\" (UniqueName: \"kubernetes.io/projected/2acd6f54-87a9-4e50-859c-42a7c8e9d098-kube-api-access-flrdg\") pod \"heat-operator-controller-manager-5f64f6f8bb-cmg9l\" (UID: \"2acd6f54-87a9-4e50-859c-42a7c8e9d098\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.194038 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjhfr\" (UniqueName: \"kubernetes.io/projected/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-kube-api-access-rjhfr\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.197310 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.201831 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.217859 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.218339 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqtzs\" (UniqueName: \"kubernetes.io/projected/b02b307a-7d4c-4d20-b388-18f6fc6ec0ee-kube-api-access-lqtzs\") pod \"glance-operator-controller-manager-5697bb5779-b7m9l\" (UID: \"b02b307a-7d4c-4d20-b388-18f6fc6ec0ee\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.234960 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2csgw\" (UniqueName: \"kubernetes.io/projected/fab26e20-8c1c-4a9e-b7c4-8d4815a1568e-kube-api-access-2csgw\") pod \"keystone-operator-controller-manager-7765d96ddf-kg2wq\" (UID: \"fab26e20-8c1c-4a9e-b7c4-8d4815a1568e\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.235024 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66shv\" (UniqueName: \"kubernetes.io/projected/1290df78-3e2c-494e-a264-c495d80c0bef-kube-api-access-66shv\") pod \"horizon-operator-controller-manager-68c6d99b8f-cggng\" (UID: \"1290df78-3e2c-494e-a264-c495d80c0bef\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.235073 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q8mz\" (UniqueName: \"kubernetes.io/projected/42d526f1-2349-4aef-80fa-a6e59f3630e9-kube-api-access-4q8mz\") pod \"ironic-operator-controller-manager-967d97867-5cwxk\" (UID: \"42d526f1-2349-4aef-80fa-a6e59f3630e9\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.235127 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpmhn\" (UniqueName: \"kubernetes.io/projected/53d61380-3a1e-45e5-a039-9d339bcdb9ea-kube-api-access-qpmhn\") pod \"manila-operator-controller-manager-5b5fd79c9c-pfq9b\" (UID: \"53d61380-3a1e-45e5-a039-9d339bcdb9ea\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.235367 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.239860 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.264611 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.309739 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.310390 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66shv\" (UniqueName: \"kubernetes.io/projected/1290df78-3e2c-494e-a264-c495d80c0bef-kube-api-access-66shv\") pod \"horizon-operator-controller-manager-68c6d99b8f-cggng\" (UID: \"1290df78-3e2c-494e-a264-c495d80c0bef\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.311644 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.335879 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-mjvsn" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.338935 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpmhn\" (UniqueName: \"kubernetes.io/projected/53d61380-3a1e-45e5-a039-9d339bcdb9ea-kube-api-access-qpmhn\") pod \"manila-operator-controller-manager-5b5fd79c9c-pfq9b\" (UID: \"53d61380-3a1e-45e5-a039-9d339bcdb9ea\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.339766 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.344558 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q8mz\" (UniqueName: \"kubernetes.io/projected/42d526f1-2349-4aef-80fa-a6e59f3630e9-kube-api-access-4q8mz\") pod \"ironic-operator-controller-manager-967d97867-5cwxk\" (UID: \"42d526f1-2349-4aef-80fa-a6e59f3630e9\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.436596 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2csgw\" (UniqueName: \"kubernetes.io/projected/fab26e20-8c1c-4a9e-b7c4-8d4815a1568e-kube-api-access-2csgw\") pod \"keystone-operator-controller-manager-7765d96ddf-kg2wq\" (UID: \"fab26e20-8c1c-4a9e-b7c4-8d4815a1568e\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.445070 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpmhn\" (UniqueName: \"kubernetes.io/projected/53d61380-3a1e-45e5-a039-9d339bcdb9ea-kube-api-access-qpmhn\") pod \"manila-operator-controller-manager-5b5fd79c9c-pfq9b\" (UID: \"53d61380-3a1e-45e5-a039-9d339bcdb9ea\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.449821 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q8mz\" (UniqueName: \"kubernetes.io/projected/42d526f1-2349-4aef-80fa-a6e59f3630e9-kube-api-access-4q8mz\") pod \"ironic-operator-controller-manager-967d97867-5cwxk\" (UID: \"42d526f1-2349-4aef-80fa-a6e59f3630e9\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.452225 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.452882 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.461459 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.463166 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.468630 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvmx7\" (UniqueName: \"kubernetes.io/projected/8ddb889c-1432-4225-a110-e0aabc0ce3bc-kube-api-access-lvmx7\") pod \"mariadb-operator-controller-manager-79c8c4686c-l8qxc\" (UID: \"8ddb889c-1432-4225-a110-e0aabc0ce3bc\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.481684 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-hmwpg" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.484054 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.508533 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.549679 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.570603 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvmx7\" (UniqueName: \"kubernetes.io/projected/8ddb889c-1432-4225-a110-e0aabc0ce3bc-kube-api-access-lvmx7\") pod \"mariadb-operator-controller-manager-79c8c4686c-l8qxc\" (UID: \"8ddb889c-1432-4225-a110-e0aabc0ce3bc\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.570740 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ph89\" (UniqueName: \"kubernetes.io/projected/1f3c9d68-2093-45a5-886e-f19c43b3e8e9-kube-api-access-8ph89\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5xgcd\" (UID: \"1f3c9d68-2093-45a5-886e-f19c43b3e8e9\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.598730 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.600396 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.611226 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-w62dn" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.617534 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.640073 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvmx7\" (UniqueName: \"kubernetes.io/projected/8ddb889c-1432-4225-a110-e0aabc0ce3bc-kube-api-access-lvmx7\") pod \"mariadb-operator-controller-manager-79c8c4686c-l8qxc\" (UID: \"8ddb889c-1432-4225-a110-e0aabc0ce3bc\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.661999 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.665498 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.665532 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.666549 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.669336 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.671793 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.673780 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-b5777" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.674599 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.676899 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ph89\" (UniqueName: \"kubernetes.io/projected/1f3c9d68-2093-45a5-886e-f19c43b3e8e9-kube-api-access-8ph89\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5xgcd\" (UID: \"1f3c9d68-2093-45a5-886e-f19c43b3e8e9\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.677097 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:15 crc kubenswrapper[4706]: E1208 19:38:15.677294 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:15 crc kubenswrapper[4706]: E1208 19:38:15.677357 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert podName:ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:16.677336949 +0000 UTC m=+979.319537952 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert") pod "infra-operator-controller-manager-78d48bff9d-qmcdh" (UID: "ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3") : secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.681997 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-29xjn" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.682908 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.690050 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.698008 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.699090 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-q2j6h" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.722038 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-sff5p" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.748361 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.749712 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ph89\" (UniqueName: \"kubernetes.io/projected/1f3c9d68-2093-45a5-886e-f19c43b3e8e9-kube-api-access-8ph89\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5xgcd\" (UID: \"1f3c9d68-2093-45a5-886e-f19c43b3e8e9\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.756556 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.757929 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.760281 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-45p8g" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.777073 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.778467 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnkjk\" (UniqueName: \"kubernetes.io/projected/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-kube-api-access-vnkjk\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.778526 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qwtg\" (UniqueName: \"kubernetes.io/projected/eedc4c77-93bc-40b9-ba3c-e7222d8828b5-kube-api-access-7qwtg\") pod \"octavia-operator-controller-manager-998648c74-5ndn8\" (UID: \"eedc4c77-93bc-40b9-ba3c-e7222d8828b5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.778576 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq59x\" (UniqueName: \"kubernetes.io/projected/6da8aab3-b971-413d-8557-cdfad2c95d2c-kube-api-access-lq59x\") pod \"nova-operator-controller-manager-697bc559fc-58rkn\" (UID: \"6da8aab3-b971-413d-8557-cdfad2c95d2c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.778630 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.779930 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.800814 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.811837 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.814088 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.816363 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.817719 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ppccd" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.828428 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.829825 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.831891 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-rbwk9" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.835336 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.843388 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.864905 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.866585 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.868900 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-527ls" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.880117 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dln88\" (UniqueName: \"kubernetes.io/projected/e9c55dcb-b67a-43f9-9a80-8c38e7f94034-kube-api-access-dln88\") pod \"swift-operator-controller-manager-9d58d64bc-7htmr\" (UID: \"e9c55dcb-b67a-43f9-9a80-8c38e7f94034\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.880172 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs78p\" (UniqueName: \"kubernetes.io/projected/ec6df814-0549-46a9-a9f3-2f94083bda52-kube-api-access-vs78p\") pod \"ovn-operator-controller-manager-b6456fdb6-vpwqp\" (UID: \"ec6df814-0549-46a9-a9f3-2f94083bda52\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.880222 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnkjk\" (UniqueName: \"kubernetes.io/projected/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-kube-api-access-vnkjk\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.880255 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qwtg\" (UniqueName: \"kubernetes.io/projected/eedc4c77-93bc-40b9-ba3c-e7222d8828b5-kube-api-access-7qwtg\") pod \"octavia-operator-controller-manager-998648c74-5ndn8\" (UID: \"eedc4c77-93bc-40b9-ba3c-e7222d8828b5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.880309 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq59x\" (UniqueName: \"kubernetes.io/projected/6da8aab3-b971-413d-8557-cdfad2c95d2c-kube-api-access-lq59x\") pod \"nova-operator-controller-manager-697bc559fc-58rkn\" (UID: \"6da8aab3-b971-413d-8557-cdfad2c95d2c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.880340 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwgk2\" (UniqueName: \"kubernetes.io/projected/22335991-2ba5-4fba-a1d1-b391e4767961-kube-api-access-dwgk2\") pod \"placement-operator-controller-manager-78f8948974-nbcmr\" (UID: \"22335991-2ba5-4fba-a1d1-b391e4767961\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.880373 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:15 crc kubenswrapper[4706]: E1208 19:38:15.880553 4706 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:15 crc kubenswrapper[4706]: E1208 19:38:15.880620 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert podName:4fca7ad0-1e55-4b7a-bee5-08521a1836a2 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:16.380600043 +0000 UTC m=+979.022801046 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f6nqtx" (UID: "4fca7ad0-1e55-4b7a-bee5-08521a1836a2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.880751 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.892974 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.903612 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.912720 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnkjk\" (UniqueName: \"kubernetes.io/projected/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-kube-api-access-vnkjk\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.913224 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qwtg\" (UniqueName: \"kubernetes.io/projected/eedc4c77-93bc-40b9-ba3c-e7222d8828b5-kube-api-access-7qwtg\") pod \"octavia-operator-controller-manager-998648c74-5ndn8\" (UID: \"eedc4c77-93bc-40b9-ba3c-e7222d8828b5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.913505 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq59x\" (UniqueName: \"kubernetes.io/projected/6da8aab3-b971-413d-8557-cdfad2c95d2c-kube-api-access-lq59x\") pod \"nova-operator-controller-manager-697bc559fc-58rkn\" (UID: \"6da8aab3-b971-413d-8557-cdfad2c95d2c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.926020 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm"] Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.947796 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.951222 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.953512 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-t8pgg" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.955919 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.979493 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.981906 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dln88\" (UniqueName: \"kubernetes.io/projected/e9c55dcb-b67a-43f9-9a80-8c38e7f94034-kube-api-access-dln88\") pod \"swift-operator-controller-manager-9d58d64bc-7htmr\" (UID: \"e9c55dcb-b67a-43f9-9a80-8c38e7f94034\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.981958 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs78p\" (UniqueName: \"kubernetes.io/projected/ec6df814-0549-46a9-a9f3-2f94083bda52-kube-api-access-vs78p\") pod \"ovn-operator-controller-manager-b6456fdb6-vpwqp\" (UID: \"ec6df814-0549-46a9-a9f3-2f94083bda52\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.982009 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwgk2\" (UniqueName: \"kubernetes.io/projected/22335991-2ba5-4fba-a1d1-b391e4767961-kube-api-access-dwgk2\") pod \"placement-operator-controller-manager-78f8948974-nbcmr\" (UID: \"22335991-2ba5-4fba-a1d1-b391e4767961\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.982061 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr6s4\" (UniqueName: \"kubernetes.io/projected/375c721a-f040-42a8-9b43-f609258063a1-kube-api-access-fr6s4\") pod \"test-operator-controller-manager-5854674fcc-qfsq9\" (UID: \"375c721a-f040-42a8-9b43-f609258063a1\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.982088 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdxhh\" (UniqueName: \"kubernetes.io/projected/ed8cb084-a66c-4fc0-afa1-068e58c80680-kube-api-access-sdxhh\") pod \"watcher-operator-controller-manager-667bd8d554-g6gls\" (UID: \"ed8cb084-a66c-4fc0-afa1-068e58c80680\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.982112 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv4s6\" (UniqueName: \"kubernetes.io/projected/beda813f-668e-45e8-8d1c-7c72caa6f294-kube-api-access-sv4s6\") pod \"telemetry-operator-controller-manager-5cbd74b459-5l5k7\" (UID: \"beda813f-668e-45e8-8d1c-7c72caa6f294\") " pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" Dec 08 19:38:15 crc kubenswrapper[4706]: I1208 19:38:15.985917 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm"] Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.036649 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs78p\" (UniqueName: \"kubernetes.io/projected/ec6df814-0549-46a9-a9f3-2f94083bda52-kube-api-access-vs78p\") pod \"ovn-operator-controller-manager-b6456fdb6-vpwqp\" (UID: \"ec6df814-0549-46a9-a9f3-2f94083bda52\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.038657 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwgk2\" (UniqueName: \"kubernetes.io/projected/22335991-2ba5-4fba-a1d1-b391e4767961-kube-api-access-dwgk2\") pod \"placement-operator-controller-manager-78f8948974-nbcmr\" (UID: \"22335991-2ba5-4fba-a1d1-b391e4767961\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.045882 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.054323 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dln88\" (UniqueName: \"kubernetes.io/projected/e9c55dcb-b67a-43f9-9a80-8c38e7f94034-kube-api-access-dln88\") pod \"swift-operator-controller-manager-9d58d64bc-7htmr\" (UID: \"e9c55dcb-b67a-43f9-9a80-8c38e7f94034\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.060144 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.072759 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2"] Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.074319 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.083362 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2"] Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.084372 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.084456 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.084490 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7bdf\" (UniqueName: \"kubernetes.io/projected/61674e93-1252-4e28-ac66-b2a701745112-kube-api-access-s7bdf\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.084569 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr6s4\" (UniqueName: \"kubernetes.io/projected/375c721a-f040-42a8-9b43-f609258063a1-kube-api-access-fr6s4\") pod \"test-operator-controller-manager-5854674fcc-qfsq9\" (UID: \"375c721a-f040-42a8-9b43-f609258063a1\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.084605 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdxhh\" (UniqueName: \"kubernetes.io/projected/ed8cb084-a66c-4fc0-afa1-068e58c80680-kube-api-access-sdxhh\") pod \"watcher-operator-controller-manager-667bd8d554-g6gls\" (UID: \"ed8cb084-a66c-4fc0-afa1-068e58c80680\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.084630 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv4s6\" (UniqueName: \"kubernetes.io/projected/beda813f-668e-45e8-8d1c-7c72caa6f294-kube-api-access-sv4s6\") pod \"telemetry-operator-controller-manager-5cbd74b459-5l5k7\" (UID: \"beda813f-668e-45e8-8d1c-7c72caa6f294\") " pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.085827 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-xbbhh" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.124197 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv4s6\" (UniqueName: \"kubernetes.io/projected/beda813f-668e-45e8-8d1c-7c72caa6f294-kube-api-access-sv4s6\") pod \"telemetry-operator-controller-manager-5cbd74b459-5l5k7\" (UID: \"beda813f-668e-45e8-8d1c-7c72caa6f294\") " pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.124479 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdxhh\" (UniqueName: \"kubernetes.io/projected/ed8cb084-a66c-4fc0-afa1-068e58c80680-kube-api-access-sdxhh\") pod \"watcher-operator-controller-manager-667bd8d554-g6gls\" (UID: \"ed8cb084-a66c-4fc0-afa1-068e58c80680\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.127607 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr6s4\" (UniqueName: \"kubernetes.io/projected/375c721a-f040-42a8-9b43-f609258063a1-kube-api-access-fr6s4\") pod \"test-operator-controller-manager-5854674fcc-qfsq9\" (UID: \"375c721a-f040-42a8-9b43-f609258063a1\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.189452 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq86b\" (UniqueName: \"kubernetes.io/projected/9cca1f0f-9037-49d5-b8da-392421a67702-kube-api-access-kq86b\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bqsg2\" (UID: \"9cca1f0f-9037-49d5-b8da-392421a67702\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.190066 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.190107 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.190129 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7bdf\" (UniqueName: \"kubernetes.io/projected/61674e93-1252-4e28-ac66-b2a701745112-kube-api-access-s7bdf\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.190521 4706 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.190690 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:16.690660563 +0000 UTC m=+979.332861566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "webhook-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.190839 4706 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.191114 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:16.690935151 +0000 UTC m=+979.333136334 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "metrics-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.211036 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7bdf\" (UniqueName: \"kubernetes.io/projected/61674e93-1252-4e28-ac66-b2a701745112-kube-api-access-s7bdf\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.269867 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.291392 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq86b\" (UniqueName: \"kubernetes.io/projected/9cca1f0f-9037-49d5-b8da-392421a67702-kube-api-access-kq86b\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bqsg2\" (UID: \"9cca1f0f-9037-49d5-b8da-392421a67702\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.306966 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.327354 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.355457 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq86b\" (UniqueName: \"kubernetes.io/projected/9cca1f0f-9037-49d5-b8da-392421a67702-kube-api-access-kq86b\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bqsg2\" (UID: \"9cca1f0f-9037-49d5-b8da-392421a67702\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.394073 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.394521 4706 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.394594 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert podName:4fca7ad0-1e55-4b7a-bee5-08521a1836a2 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:17.394573286 +0000 UTC m=+980.036774289 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f6nqtx" (UID: "4fca7ad0-1e55-4b7a-bee5-08521a1836a2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.561128 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.568144 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw"] Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.570962 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.608110 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.700751 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.700995 4706 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.701492 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:17.701462055 +0000 UTC m=+980.343663238 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "metrics-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.701527 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.701605 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert podName:ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:18.701583678 +0000 UTC m=+981.343784691 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert") pod "infra-operator-controller-manager-78d48bff9d-qmcdh" (UID: "ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3") : secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.701367 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:16 crc kubenswrapper[4706]: I1208 19:38:16.702045 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.702284 4706 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 19:38:16 crc kubenswrapper[4706]: E1208 19:38:16.702327 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:17.702315419 +0000 UTC m=+980.344516422 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "webhook-server-cert" not found Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.084354 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.104329 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl"] Dec 08 19:38:17 crc kubenswrapper[4706]: W1208 19:38:17.104871 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f144ce9_7cc0_41f3_84ea_654579d32c0b.slice/crio-77bccdc69393316f553470b369388ceae12babd76066437b73f577d3b244ce50 WatchSource:0}: Error finding container 77bccdc69393316f553470b369388ceae12babd76066437b73f577d3b244ce50: Status 404 returned error can't find the container with id 77bccdc69393316f553470b369388ceae12babd76066437b73f577d3b244ce50 Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.144085 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.222119 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.360418 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.377122 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.387649 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.412835 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.415134 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.415354 4706 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.415445 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert podName:4fca7ad0-1e55-4b7a-bee5-08521a1836a2 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:19.415415487 +0000 UTC m=+982.057616490 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f6nqtx" (UID: "4fca7ad0-1e55-4b7a-bee5-08521a1836a2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:17 crc kubenswrapper[4706]: W1208 19:38:17.424032 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ddb889c_1432_4225_a110_e0aabc0ce3bc.slice/crio-7768ce4f796a71ce1044a9dbe6d40cfce07461a3395c9092e96a91ce14d3be0d WatchSource:0}: Error finding container 7768ce4f796a71ce1044a9dbe6d40cfce07461a3395c9092e96a91ce14d3be0d: Status 404 returned error can't find the container with id 7768ce4f796a71ce1044a9dbe6d40cfce07461a3395c9092e96a91ce14d3be0d Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.426499 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.442128 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.448318 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.470992 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" event={"ID":"1290df78-3e2c-494e-a264-c495d80c0bef","Type":"ContainerStarted","Data":"234ca2764a0a4bade38bd3c48dd02b8c2d14ce8ad0028ed4dff1f4f44e0ea363"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.472620 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" event={"ID":"ec6df814-0549-46a9-a9f3-2f94083bda52","Type":"ContainerStarted","Data":"27ccacfe2fb536f7c1bef5cfe29076cac7bfadd6c4b7fe0117bb1d532c655d41"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.473846 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" event={"ID":"b02b307a-7d4c-4d20-b388-18f6fc6ec0ee","Type":"ContainerStarted","Data":"e50e211bdd64fc75d7190fa85f9fa3ee0a1c9ae4a72d92c09aff2fe77529f1a2"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.475411 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" event={"ID":"1f3c9d68-2093-45a5-886e-f19c43b3e8e9","Type":"ContainerStarted","Data":"27528fa797a749883372151f250fb09acceac7745e9ce452394176ddb7fe6b9a"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.477150 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" event={"ID":"fab26e20-8c1c-4a9e-b7c4-8d4815a1568e","Type":"ContainerStarted","Data":"cb959ea89e7f093c9aafda9736863ab5ba135e901cd89a3d31bbe97fb39b7149"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.478064 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" event={"ID":"4c42f8b9-a85c-4a45-a8f8-4a43e561a14e","Type":"ContainerStarted","Data":"ce39e2696ff9346bf0d4c889dd26f581b7adab603ee7e448a0dc2734337bfb76"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.479326 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" event={"ID":"42d526f1-2349-4aef-80fa-a6e59f3630e9","Type":"ContainerStarted","Data":"e8f19ebc156f9b534ffcf40b84b33a24e1ff73381f881c8f2765dbe96169c0e6"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.509527 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" event={"ID":"0f144ce9-7cc0-41f3-84ea-654579d32c0b","Type":"ContainerStarted","Data":"77bccdc69393316f553470b369388ceae12babd76066437b73f577d3b244ce50"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.511441 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" event={"ID":"8ddb889c-1432-4225-a110-e0aabc0ce3bc","Type":"ContainerStarted","Data":"7768ce4f796a71ce1044a9dbe6d40cfce07461a3395c9092e96a91ce14d3be0d"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.514342 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" event={"ID":"eedc4c77-93bc-40b9-ba3c-e7222d8828b5","Type":"ContainerStarted","Data":"a8435e2c68c599a8ec468ac5fea35654a93b4a9c2e97f1719c49cedcb4276e92"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.516123 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" event={"ID":"2acd6f54-87a9-4e50-859c-42a7c8e9d098","Type":"ContainerStarted","Data":"3f68e660a3b7e57822e061dbf5f73672f40ffd13491ab15fc926b8acc83b89f4"} Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.518487 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" event={"ID":"e5899e67-6618-43f5-8454-fddfae494e4a","Type":"ContainerStarted","Data":"97bff6c65c91b0bf7a267c800d04665b3bfba182bfa44f448f49f05618ec6c5b"} Dec 08 19:38:17 crc kubenswrapper[4706]: W1208 19:38:17.522937 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53d61380_3a1e_45e5_a039_9d339bcdb9ea.slice/crio-8994e7c182f672380b151520f0226a859fbcce17f9026d090b8920c5d025485b WatchSource:0}: Error finding container 8994e7c182f672380b151520f0226a859fbcce17f9026d090b8920c5d025485b: Status 404 returned error can't find the container with id 8994e7c182f672380b151520f0226a859fbcce17f9026d090b8920c5d025485b Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.525831 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qpmhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-pfq9b_openstack-operators(53d61380-3a1e-45e5-a039-9d339bcdb9ea): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.529511 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qpmhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-pfq9b_openstack-operators(53d61380-3a1e-45e5-a039-9d339bcdb9ea): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.530989 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" podUID="53d61380-3a1e-45e5-a039-9d339bcdb9ea" Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.536011 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.668020 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr"] Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.733014 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dwgk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-nbcmr_openstack-operators(22335991-2ba5-4fba-a1d1-b391e4767961): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.733578 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.733762 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.733961 4706 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.734046 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:19.734022531 +0000 UTC m=+982.376223524 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "webhook-server-cert" not found Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.734112 4706 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.734138 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:19.734129884 +0000 UTC m=+982.376330887 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "metrics-server-cert" not found Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.737140 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dwgk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-nbcmr_openstack-operators(22335991-2ba5-4fba-a1d1-b391e4767961): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.737504 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lq59x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-58rkn_openstack-operators(6da8aab3-b971-413d-8557-cdfad2c95d2c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.737763 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn"] Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.742056 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" podUID="22335991-2ba5-4fba-a1d1-b391e4767961" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.747947 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lq59x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-58rkn_openstack-operators(6da8aab3-b971-413d-8557-cdfad2c95d2c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.750594 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" podUID="6da8aab3-b971-413d-8557-cdfad2c95d2c" Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.760139 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr"] Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.811957 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7"] Dec 08 19:38:17 crc kubenswrapper[4706]: W1208 19:38:17.844216 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded8cb084_a66c_4fc0_afa1_068e58c80680.slice/crio-0563ce094154f7886022ecd36224e0e3ef723121f8eb6a8dab2a9aac16d23cf9 WatchSource:0}: Error finding container 0563ce094154f7886022ecd36224e0e3ef723121f8eb6a8dab2a9aac16d23cf9: Status 404 returned error can't find the container with id 0563ce094154f7886022ecd36224e0e3ef723121f8eb6a8dab2a9aac16d23cf9 Dec 08 19:38:17 crc kubenswrapper[4706]: W1208 19:38:17.848543 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbeda813f_668e_45e8_8d1c_7c72caa6f294.slice/crio-46a650715bbb3e00bc35e1434d98e9d6f4f9d065cafbddca1a6be353b7f15a45 WatchSource:0}: Error finding container 46a650715bbb3e00bc35e1434d98e9d6f4f9d065cafbddca1a6be353b7f15a45: Status 404 returned error can't find the container with id 46a650715bbb3e00bc35e1434d98e9d6f4f9d065cafbddca1a6be353b7f15a45 Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.856823 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9"] Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.867151 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fr6s4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-qfsq9_openstack-operators(375c721a-f040-42a8-9b43-f609258063a1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.869056 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls"] Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.871175 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fr6s4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-qfsq9_openstack-operators(375c721a-f040-42a8-9b43-f609258063a1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.872896 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" podUID="375c721a-f040-42a8-9b43-f609258063a1" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.875771 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.51:5001/openstack-k8s-operators/telemetry-operator:d3ea47b1122f22fdda4bc30dd95b8db90981973f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sv4s6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5cbd74b459-5l5k7_openstack-operators(beda813f-668e-45e8-8d1c-7c72caa6f294): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.878539 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sv4s6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5cbd74b459-5l5k7_openstack-operators(beda813f-668e-45e8-8d1c-7c72caa6f294): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.880210 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" podUID="beda813f-668e-45e8-8d1c-7c72caa6f294" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.887220 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kq86b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-bqsg2_openstack-operators(9cca1f0f-9037-49d5-b8da-392421a67702): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:38:17 crc kubenswrapper[4706]: E1208 19:38:17.889545 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" podUID="9cca1f0f-9037-49d5-b8da-392421a67702" Dec 08 19:38:17 crc kubenswrapper[4706]: I1208 19:38:17.896232 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2"] Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.567490 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" event={"ID":"22335991-2ba5-4fba-a1d1-b391e4767961","Type":"ContainerStarted","Data":"3beeef95ece349109e1e5efe77e8a5f57287aa2066e060d9862ada74795ab503"} Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.570253 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" event={"ID":"e9c55dcb-b67a-43f9-9a80-8c38e7f94034","Type":"ContainerStarted","Data":"b116d205b1b53430acd8711f7f43a81fbc31babb55600535107b21c57213e9e9"} Dec 08 19:38:18 crc kubenswrapper[4706]: E1208 19:38:18.573903 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" podUID="22335991-2ba5-4fba-a1d1-b391e4767961" Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.578516 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" event={"ID":"beda813f-668e-45e8-8d1c-7c72caa6f294","Type":"ContainerStarted","Data":"46a650715bbb3e00bc35e1434d98e9d6f4f9d065cafbddca1a6be353b7f15a45"} Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.580365 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" event={"ID":"53d61380-3a1e-45e5-a039-9d339bcdb9ea","Type":"ContainerStarted","Data":"8994e7c182f672380b151520f0226a859fbcce17f9026d090b8920c5d025485b"} Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.581539 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" event={"ID":"375c721a-f040-42a8-9b43-f609258063a1","Type":"ContainerStarted","Data":"a2b3d8a0fec5419fe8ef53f2f5dc67c88a099b2d52cda630828047ac967ac3af"} Dec 08 19:38:18 crc kubenswrapper[4706]: E1208 19:38:18.585033 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.51:5001/openstack-k8s-operators/telemetry-operator:d3ea47b1122f22fdda4bc30dd95b8db90981973f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" podUID="beda813f-668e-45e8-8d1c-7c72caa6f294" Dec 08 19:38:18 crc kubenswrapper[4706]: E1208 19:38:18.589939 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" podUID="53d61380-3a1e-45e5-a039-9d339bcdb9ea" Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.591550 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" event={"ID":"6da8aab3-b971-413d-8557-cdfad2c95d2c","Type":"ContainerStarted","Data":"56a980c0b6b867c55c34f0f190c38824124b8032603b2efcfb9b8df9d292a9c3"} Dec 08 19:38:18 crc kubenswrapper[4706]: E1208 19:38:18.594613 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" podUID="375c721a-f040-42a8-9b43-f609258063a1" Dec 08 19:38:18 crc kubenswrapper[4706]: E1208 19:38:18.597501 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" podUID="6da8aab3-b971-413d-8557-cdfad2c95d2c" Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.597706 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" event={"ID":"9cca1f0f-9037-49d5-b8da-392421a67702","Type":"ContainerStarted","Data":"facd77ecbf34585dc08db66ec51e0508fac0b789e226d8bf29b3adf7187b9f33"} Dec 08 19:38:18 crc kubenswrapper[4706]: E1208 19:38:18.599983 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" podUID="9cca1f0f-9037-49d5-b8da-392421a67702" Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.610125 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" event={"ID":"ed8cb084-a66c-4fc0-afa1-068e58c80680","Type":"ContainerStarted","Data":"0563ce094154f7886022ecd36224e0e3ef723121f8eb6a8dab2a9aac16d23cf9"} Dec 08 19:38:18 crc kubenswrapper[4706]: I1208 19:38:18.766160 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:18 crc kubenswrapper[4706]: E1208 19:38:18.766596 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:18 crc kubenswrapper[4706]: E1208 19:38:18.766811 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert podName:ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:22.766754584 +0000 UTC m=+985.408955587 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert") pod "infra-operator-controller-manager-78d48bff9d-qmcdh" (UID: "ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3") : secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:19 crc kubenswrapper[4706]: I1208 19:38:19.478496 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.478733 4706 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.478800 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert podName:4fca7ad0-1e55-4b7a-bee5-08521a1836a2 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:23.478779562 +0000 UTC m=+986.120980565 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f6nqtx" (UID: "4fca7ad0-1e55-4b7a-bee5-08521a1836a2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.635456 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" podUID="9cca1f0f-9037-49d5-b8da-392421a67702" Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.635693 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" podUID="6da8aab3-b971-413d-8557-cdfad2c95d2c" Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.635897 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" podUID="375c721a-f040-42a8-9b43-f609258063a1" Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.646736 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" podUID="53d61380-3a1e-45e5-a039-9d339bcdb9ea" Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.646674 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" podUID="22335991-2ba5-4fba-a1d1-b391e4767961" Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.656090 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.51:5001/openstack-k8s-operators/telemetry-operator:d3ea47b1122f22fdda4bc30dd95b8db90981973f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" podUID="beda813f-668e-45e8-8d1c-7c72caa6f294" Dec 08 19:38:19 crc kubenswrapper[4706]: I1208 19:38:19.787022 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:19 crc kubenswrapper[4706]: I1208 19:38:19.787148 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.787944 4706 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.788016 4706 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.788295 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:23.788067769 +0000 UTC m=+986.430268932 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "metrics-server-cert" not found Dec 08 19:38:19 crc kubenswrapper[4706]: E1208 19:38:19.788431 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:23.788417699 +0000 UTC m=+986.430618862 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "webhook-server-cert" not found Dec 08 19:38:22 crc kubenswrapper[4706]: I1208 19:38:22.843052 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:22 crc kubenswrapper[4706]: E1208 19:38:22.843254 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:22 crc kubenswrapper[4706]: E1208 19:38:22.843810 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert podName:ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:30.843751439 +0000 UTC m=+993.485952442 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert") pod "infra-operator-controller-manager-78d48bff9d-qmcdh" (UID: "ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3") : secret "infra-operator-webhook-server-cert" not found Dec 08 19:38:23 crc kubenswrapper[4706]: I1208 19:38:23.558724 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:23 crc kubenswrapper[4706]: E1208 19:38:23.558958 4706 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:23 crc kubenswrapper[4706]: E1208 19:38:23.559110 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert podName:4fca7ad0-1e55-4b7a-bee5-08521a1836a2 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:31.559087172 +0000 UTC m=+994.201288175 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f6nqtx" (UID: "4fca7ad0-1e55-4b7a-bee5-08521a1836a2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 19:38:23 crc kubenswrapper[4706]: I1208 19:38:23.862751 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:23 crc kubenswrapper[4706]: I1208 19:38:23.862942 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:23 crc kubenswrapper[4706]: E1208 19:38:23.863127 4706 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 19:38:23 crc kubenswrapper[4706]: E1208 19:38:23.863205 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:31.863185751 +0000 UTC m=+994.505386754 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "webhook-server-cert" not found Dec 08 19:38:23 crc kubenswrapper[4706]: E1208 19:38:23.864098 4706 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 19:38:23 crc kubenswrapper[4706]: E1208 19:38:23.864313 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:31.864285263 +0000 UTC m=+994.506486276 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "metrics-server-cert" not found Dec 08 19:38:29 crc kubenswrapper[4706]: E1208 19:38:29.437111 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 08 19:38:29 crc kubenswrapper[4706]: E1208 19:38:29.438209 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66shv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-cggng_openstack-operators(1290df78-3e2c-494e-a264-c495d80c0bef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:30 crc kubenswrapper[4706]: E1208 19:38:30.289835 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 08 19:38:30 crc kubenswrapper[4706]: E1208 19:38:30.290595 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vs78p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-vpwqp_openstack-operators(ec6df814-0549-46a9-a9f3-2f94083bda52): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:30 crc kubenswrapper[4706]: E1208 19:38:30.873833 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 08 19:38:30 crc kubenswrapper[4706]: E1208 19:38:30.874083 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gzjxw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6c677c69b-j2mjl_openstack-operators(e5899e67-6618-43f5-8454-fddfae494e4a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:30 crc kubenswrapper[4706]: I1208 19:38:30.929866 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:30 crc kubenswrapper[4706]: I1208 19:38:30.937176 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-qmcdh\" (UID: \"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:30 crc kubenswrapper[4706]: I1208 19:38:30.968224 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:38:31 crc kubenswrapper[4706]: I1208 19:38:31.652743 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:31 crc kubenswrapper[4706]: I1208 19:38:31.658792 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fca7ad0-1e55-4b7a-bee5-08521a1836a2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f6nqtx\" (UID: \"4fca7ad0-1e55-4b7a-bee5-08521a1836a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:31 crc kubenswrapper[4706]: I1208 19:38:31.908656 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:38:31 crc kubenswrapper[4706]: I1208 19:38:31.958145 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:31 crc kubenswrapper[4706]: I1208 19:38:31.958238 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:31 crc kubenswrapper[4706]: E1208 19:38:31.958421 4706 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 19:38:31 crc kubenswrapper[4706]: E1208 19:38:31.958521 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:47.958496273 +0000 UTC m=+1010.600697286 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "webhook-server-cert" not found Dec 08 19:38:31 crc kubenswrapper[4706]: E1208 19:38:31.958587 4706 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 19:38:31 crc kubenswrapper[4706]: E1208 19:38:31.958703 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs podName:61674e93-1252-4e28-ac66-b2a701745112 nodeName:}" failed. No retries permitted until 2025-12-08 19:38:47.958676938 +0000 UTC m=+1010.600877951 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs") pod "openstack-operator-controller-manager-b89665ccc-lvggm" (UID: "61674e93-1252-4e28-ac66-b2a701745112") : secret "metrics-server-cert" not found Dec 08 19:38:32 crc kubenswrapper[4706]: E1208 19:38:32.294941 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 08 19:38:32 crc kubenswrapper[4706]: E1208 19:38:32.295170 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lvmx7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-l8qxc_openstack-operators(8ddb889c-1432-4225-a110-e0aabc0ce3bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:33 crc kubenswrapper[4706]: E1208 19:38:33.010835 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 08 19:38:33 crc kubenswrapper[4706]: E1208 19:38:33.011675 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2csgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-kg2wq_openstack-operators(fab26e20-8c1c-4a9e-b7c4-8d4815a1568e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:33 crc kubenswrapper[4706]: I1208 19:38:33.679011 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx"] Dec 08 19:38:33 crc kubenswrapper[4706]: I1208 19:38:33.741582 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh"] Dec 08 19:38:33 crc kubenswrapper[4706]: I1208 19:38:33.768861 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" event={"ID":"b02b307a-7d4c-4d20-b388-18f6fc6ec0ee","Type":"ContainerStarted","Data":"2c220a7fd046609f9a382b371baa8f078eca8bd16f4354862f12f63d1237f96b"} Dec 08 19:38:33 crc kubenswrapper[4706]: I1208 19:38:33.776866 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" event={"ID":"0f144ce9-7cc0-41f3-84ea-654579d32c0b","Type":"ContainerStarted","Data":"542ff704949ae5bd43b874a3802d406b52c2dbc12b878504bedf9ad9a11b0523"} Dec 08 19:38:33 crc kubenswrapper[4706]: I1208 19:38:33.851314 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" event={"ID":"eedc4c77-93bc-40b9-ba3c-e7222d8828b5","Type":"ContainerStarted","Data":"9099731a1647a4072a4e6951a7c5f89f977f06f2d7259798a7c1737fbe8e038f"} Dec 08 19:38:33 crc kubenswrapper[4706]: I1208 19:38:33.864608 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" event={"ID":"e9c55dcb-b67a-43f9-9a80-8c38e7f94034","Type":"ContainerStarted","Data":"7c4ad9b2e4d906911af2959fb9b24cc7fd4bf3ea1cf36a4dacd4ea7d8e064ec4"} Dec 08 19:38:33 crc kubenswrapper[4706]: I1208 19:38:33.872010 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" event={"ID":"42d526f1-2349-4aef-80fa-a6e59f3630e9","Type":"ContainerStarted","Data":"d92d732c4524386a22042f660030a98952254704dbbcdac88dc451a2c060803d"} Dec 08 19:38:34 crc kubenswrapper[4706]: I1208 19:38:34.924537 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" event={"ID":"2acd6f54-87a9-4e50-859c-42a7c8e9d098","Type":"ContainerStarted","Data":"f84e1ac1eb493244080690c6c29409f18d9f728e681c898c949b0708de8bbf23"} Dec 08 19:38:34 crc kubenswrapper[4706]: I1208 19:38:34.952512 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" event={"ID":"4c42f8b9-a85c-4a45-a8f8-4a43e561a14e","Type":"ContainerStarted","Data":"9603e2036eb1440c203ae73e86ed893b9bd58e169649557eded672c8a8ebe2eb"} Dec 08 19:38:34 crc kubenswrapper[4706]: I1208 19:38:34.965393 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" event={"ID":"4fca7ad0-1e55-4b7a-bee5-08521a1836a2","Type":"ContainerStarted","Data":"3b5df500c02452f62dcaf47a24a3a487bd261da16a7b2793d334a5986c70c27b"} Dec 08 19:38:34 crc kubenswrapper[4706]: I1208 19:38:34.982705 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" event={"ID":"ed8cb084-a66c-4fc0-afa1-068e58c80680","Type":"ContainerStarted","Data":"418c1c07857214e6358895588d202745fa5b993b9e7d00268feabe42a3b8e819"} Dec 08 19:38:34 crc kubenswrapper[4706]: I1208 19:38:34.994207 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" event={"ID":"1f3c9d68-2093-45a5-886e-f19c43b3e8e9","Type":"ContainerStarted","Data":"2e6b902248612c4207f547c3c603ec70ef8a4707e475aa0a2e99ceed2a327cff"} Dec 08 19:38:35 crc kubenswrapper[4706]: I1208 19:38:35.836552 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:38:35 crc kubenswrapper[4706]: I1208 19:38:35.837140 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:38:36 crc kubenswrapper[4706]: I1208 19:38:36.021341 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" event={"ID":"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3","Type":"ContainerStarted","Data":"436e9db37977a1b480511c0bf8c233d02b14eba744cf449dce9afbd09922a73f"} Dec 08 19:38:48 crc kubenswrapper[4706]: I1208 19:38:48.031188 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:48 crc kubenswrapper[4706]: I1208 19:38:48.032103 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:48 crc kubenswrapper[4706]: I1208 19:38:48.038318 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-webhook-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:48 crc kubenswrapper[4706]: I1208 19:38:48.038367 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61674e93-1252-4e28-ac66-b2a701745112-metrics-certs\") pod \"openstack-operator-controller-manager-b89665ccc-lvggm\" (UID: \"61674e93-1252-4e28-ac66-b2a701745112\") " pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:48 crc kubenswrapper[4706]: I1208 19:38:48.093637 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:38:54 crc kubenswrapper[4706]: E1208 19:38:54.198106 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 08 19:38:54 crc kubenswrapper[4706]: E1208 19:38:54.199356 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dwgk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-nbcmr_openstack-operators(22335991-2ba5-4fba-a1d1-b391e4767961): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.072363 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.072653 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.072950 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4jcvs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-qtnlw_openstack-operators(4c42f8b9-a85c-4a45-a8f8-4a43e561a14e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.073094 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ssbxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-7fvjs_openstack-operators(0f144ce9-7cc0-41f3-84ea-654579d32c0b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.073139 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.073367 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lqtzs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-b7m9l_openstack-operators(b02b307a-7d4c-4d20-b388-18f6fc6ec0ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.074243 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" podUID="4c42f8b9-a85c-4a45-a8f8-4a43e561a14e" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.074285 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" podUID="0f144ce9-7cc0-41f3-84ea-654579d32c0b" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.074709 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" podUID="b02b307a-7d4c-4d20-b388-18f6fc6ec0ee" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.078493 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.078746 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2csgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-kg2wq_openstack-operators(fab26e20-8c1c-4a9e-b7c4-8d4815a1568e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.080565 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" podUID="fab26e20-8c1c-4a9e-b7c4-8d4815a1568e" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.088285 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.088514 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lq59x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-58rkn_openstack-operators(6da8aab3-b971-413d-8557-cdfad2c95d2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:57 crc kubenswrapper[4706]: I1208 19:38:57.266965 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" Dec 08 19:38:57 crc kubenswrapper[4706]: I1208 19:38:57.267013 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.268315 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" podUID="b02b307a-7d4c-4d20-b388-18f6fc6ec0ee" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.268497 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" podUID="4c42f8b9-a85c-4a45-a8f8-4a43e561a14e" Dec 08 19:38:57 crc kubenswrapper[4706]: I1208 19:38:57.268975 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.270298 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" podUID="0f144ce9-7cc0-41f3-84ea-654579d32c0b" Dec 08 19:38:57 crc kubenswrapper[4706]: I1208 19:38:57.270970 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.788668 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.788919 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.789221 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sdxhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-g6gls_openstack-operators(ed8cb084-a66c-4fc0-afa1-068e58c80680): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.789291 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7qwtg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-5ndn8_openstack-operators(eedc4c77-93bc-40b9-ba3c-e7222d8828b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.790534 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" podUID="eedc4c77-93bc-40b9-ba3c-e7222d8828b5" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.790590 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" podUID="ed8cb084-a66c-4fc0-afa1-068e58c80680" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.807364 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48" Dec 08 19:38:57 crc kubenswrapper[4706]: E1208 19:38:57.807882 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnkjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-84b575879f6nqtx_openstack-operators(4fca7ad0-1e55-4b7a-bee5-08521a1836a2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:58 crc kubenswrapper[4706]: I1208 19:38:58.273982 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.275601 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" podUID="eedc4c77-93bc-40b9-ba3c-e7222d8828b5" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.275719 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" podUID="ed8cb084-a66c-4fc0-afa1-068e58c80680" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.276078 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" podUID="0f144ce9-7cc0-41f3-84ea-654579d32c0b" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.276482 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" podUID="b02b307a-7d4c-4d20-b388-18f6fc6ec0ee" Dec 08 19:38:58 crc kubenswrapper[4706]: I1208 19:38:58.278204 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.463242 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.463480 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8ph89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-5xgcd_openstack-operators(1f3c9d68-2093-45a5-886e-f19c43b3e8e9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.463663 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.463739 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4q8mz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-5cwxk_openstack-operators(42d526f1-2349-4aef-80fa-a6e59f3630e9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.465393 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" podUID="42d526f1-2349-4aef-80fa-a6e59f3630e9" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.465420 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" podUID="1f3c9d68-2093-45a5-886e-f19c43b3e8e9" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.473992 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:ccc60d56d8efc2e91a7d8a7131eb7e06c189c32247f2a819818c084ba2e2f2ab" Dec 08 19:38:58 crc kubenswrapper[4706]: E1208 19:38:58.474307 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:ccc60d56d8efc2e91a7d8a7131eb7e06c189c32247f2a819818c084ba2e2f2ab,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rjhfr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-78d48bff9d-qmcdh_openstack-operators(ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:59 crc kubenswrapper[4706]: E1208 19:38:59.155502 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 08 19:38:59 crc kubenswrapper[4706]: E1208 19:38:59.155758 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kq86b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-bqsg2_openstack-operators(9cca1f0f-9037-49d5-b8da-392421a67702): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:38:59 crc kubenswrapper[4706]: E1208 19:38:59.156959 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" podUID="9cca1f0f-9037-49d5-b8da-392421a67702" Dec 08 19:38:59 crc kubenswrapper[4706]: I1208 19:38:59.322872 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" Dec 08 19:38:59 crc kubenswrapper[4706]: I1208 19:38:59.323359 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" Dec 08 19:38:59 crc kubenswrapper[4706]: I1208 19:38:59.325689 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" Dec 08 19:38:59 crc kubenswrapper[4706]: I1208 19:38:59.328627 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" Dec 08 19:38:59 crc kubenswrapper[4706]: I1208 19:38:59.698979 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm"] Dec 08 19:38:59 crc kubenswrapper[4706]: W1208 19:38:59.750137 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61674e93_1252_4e28_ac66_b2a701745112.slice/crio-c47d8c863b3ad3f05cd45c78be8a3c81db4791d8f3e8dfb275a65b6eeddb17a6 WatchSource:0}: Error finding container c47d8c863b3ad3f05cd45c78be8a3c81db4791d8f3e8dfb275a65b6eeddb17a6: Status 404 returned error can't find the container with id c47d8c863b3ad3f05cd45c78be8a3c81db4791d8f3e8dfb275a65b6eeddb17a6 Dec 08 19:38:59 crc kubenswrapper[4706]: E1208 19:38:59.801584 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" podUID="8ddb889c-1432-4225-a110-e0aabc0ce3bc" Dec 08 19:38:59 crc kubenswrapper[4706]: E1208 19:38:59.831946 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" podUID="ec6df814-0549-46a9-a9f3-2f94083bda52" Dec 08 19:38:59 crc kubenswrapper[4706]: E1208 19:38:59.918009 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" podUID="1290df78-3e2c-494e-a264-c495d80c0bef" Dec 08 19:39:00 crc kubenswrapper[4706]: E1208 19:39:00.000619 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" podUID="4fca7ad0-1e55-4b7a-bee5-08521a1836a2" Dec 08 19:39:00 crc kubenswrapper[4706]: E1208 19:39:00.214618 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" podUID="e5899e67-6618-43f5-8454-fddfae494e4a" Dec 08 19:39:00 crc kubenswrapper[4706]: E1208 19:39:00.249880 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" podUID="22335991-2ba5-4fba-a1d1-b391e4767961" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.367539 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" event={"ID":"1290df78-3e2c-494e-a264-c495d80c0bef","Type":"ContainerStarted","Data":"0bf05b0b501ba61ac695984d9e3edf78a5baab50e8f48cb9d2deffa63fcea4f6"} Dec 08 19:39:00 crc kubenswrapper[4706]: E1208 19:39:00.369082 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" podUID="ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.377520 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" event={"ID":"ec6df814-0549-46a9-a9f3-2f94083bda52","Type":"ContainerStarted","Data":"7a34de77260c11aff1156db0ce7fee29bff480f04c6004f9797ccf7f8a9ea75e"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.388381 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" event={"ID":"42d526f1-2349-4aef-80fa-a6e59f3630e9","Type":"ContainerStarted","Data":"93a5d1c45dc06c4c54197ad35ea75ff2ac35be1a885b4d7c5f47fb238880c30c"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.393952 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" event={"ID":"22335991-2ba5-4fba-a1d1-b391e4767961","Type":"ContainerStarted","Data":"f28f0f64f02ee00088f25f3cdd45a77bb6fd76fc813e55a7297f4f4a628070b9"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.396540 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" event={"ID":"4fca7ad0-1e55-4b7a-bee5-08521a1836a2","Type":"ContainerStarted","Data":"69b122c13732ebe28e7f3fff9217f70bb8e714509957aaf22d8033081092c978"} Dec 08 19:39:00 crc kubenswrapper[4706]: E1208 19:39:00.396621 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" podUID="22335991-2ba5-4fba-a1d1-b391e4767961" Dec 08 19:39:00 crc kubenswrapper[4706]: E1208 19:39:00.397905 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" podUID="4fca7ad0-1e55-4b7a-bee5-08521a1836a2" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.398943 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" event={"ID":"53d61380-3a1e-45e5-a039-9d339bcdb9ea","Type":"ContainerStarted","Data":"dbe7cb4f4141199b8c46e3e622389352518c5cae33607de990b558a79f6ee578"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.406673 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" event={"ID":"e5899e67-6618-43f5-8454-fddfae494e4a","Type":"ContainerStarted","Data":"d3c36c7f7d5c86ffc4bbb02aab243edb9e059ea64391944b67e590e4ef5d9882"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.439184 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" event={"ID":"e9c55dcb-b67a-43f9-9a80-8c38e7f94034","Type":"ContainerStarted","Data":"9b285853befae4280a0e89f601dc06ee02bfecc8260fac11cfaedf5ab5e913bd"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.440854 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.454516 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" event={"ID":"8ddb889c-1432-4225-a110-e0aabc0ce3bc","Type":"ContainerStarted","Data":"3253bc927ea36673bfae93521ef85352e416c31dd0674637bc48ff30dce6231e"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.455607 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.497317 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" event={"ID":"2acd6f54-87a9-4e50-859c-42a7c8e9d098","Type":"ContainerStarted","Data":"b7111cbaf2c06f8fa997652c5344d3d3e6916b8606e4e866c9e6824b4a498598"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.497968 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.502435 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5cwxk" podStartSLOduration=29.818095051 podStartE2EDuration="45.502405382s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.392251495 +0000 UTC m=+980.034452498" lastFinishedPulling="2025-12-08 19:38:33.076561826 +0000 UTC m=+995.718762829" observedRunningTime="2025-12-08 19:39:00.487170716 +0000 UTC m=+1023.129371719" watchObservedRunningTime="2025-12-08 19:39:00.502405382 +0000 UTC m=+1023.144606385" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.507487 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.514858 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" event={"ID":"fab26e20-8c1c-4a9e-b7c4-8d4815a1568e","Type":"ContainerStarted","Data":"50cc0b60655b6d07f1716e1ca2049b653b42b120abb0b5b807d9a0d6d0043856"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.552872 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" event={"ID":"beda813f-668e-45e8-8d1c-7c72caa6f294","Type":"ContainerStarted","Data":"e229a35abb37001ecf721259c06321fb7eda90d01956a045da0ceed86a90e1d4"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.575702 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" event={"ID":"375c721a-f040-42a8-9b43-f609258063a1","Type":"ContainerStarted","Data":"08095fbcb60cd6375cb608d31336e0332385c66bef870523fc02ef227a61a2fc"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.589936 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" event={"ID":"61674e93-1252-4e28-ac66-b2a701745112","Type":"ContainerStarted","Data":"c47d8c863b3ad3f05cd45c78be8a3c81db4791d8f3e8dfb275a65b6eeddb17a6"} Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.625393 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cmg9l" podStartSLOduration=4.568177752 podStartE2EDuration="46.625363939s" podCreationTimestamp="2025-12-08 19:38:14 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.150767677 +0000 UTC m=+979.792968670" lastFinishedPulling="2025-12-08 19:38:59.207953854 +0000 UTC m=+1021.850154857" observedRunningTime="2025-12-08 19:39:00.60197421 +0000 UTC m=+1023.244175213" watchObservedRunningTime="2025-12-08 19:39:00.625363939 +0000 UTC m=+1023.267564942" Dec 08 19:39:00 crc kubenswrapper[4706]: I1208 19:39:00.640831 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-7htmr" podStartSLOduration=4.076620887 podStartE2EDuration="45.640806701s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.66614837 +0000 UTC m=+980.308349373" lastFinishedPulling="2025-12-08 19:38:59.230334174 +0000 UTC m=+1021.872535187" observedRunningTime="2025-12-08 19:39:00.636451947 +0000 UTC m=+1023.278652950" watchObservedRunningTime="2025-12-08 19:39:00.640806701 +0000 UTC m=+1023.283007704" Dec 08 19:39:00 crc kubenswrapper[4706]: E1208 19:39:00.865220 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" podUID="6da8aab3-b971-413d-8557-cdfad2c95d2c" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.704781 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.704856 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" event={"ID":"375c721a-f040-42a8-9b43-f609258063a1","Type":"ContainerStarted","Data":"74107808c83e490500e8980f2e58fc71c793a3b74238f28c0f6adc2df57b35d6"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.704897 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" event={"ID":"6da8aab3-b971-413d-8557-cdfad2c95d2c","Type":"ContainerStarted","Data":"301ccd3ded5390e45a8f79ede5923aaff9e432a5f2aa29307b908209093faffb"} Dec 08 19:39:01 crc kubenswrapper[4706]: E1208 19:39:01.707943 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" podUID="6da8aab3-b971-413d-8557-cdfad2c95d2c" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.712936 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" event={"ID":"61674e93-1252-4e28-ac66-b2a701745112","Type":"ContainerStarted","Data":"10c8e3fb2da9e816b7cca0c069ffcecc4dbbb3a543f8add128eba6742bf6178a"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.713752 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.745003 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" event={"ID":"e5899e67-6618-43f5-8454-fddfae494e4a","Type":"ContainerStarted","Data":"a8c612a2b3f51d6b1b489baed9fea53d94461ef735a1550386eb8a9e426e38e1"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.746191 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.753502 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" event={"ID":"fab26e20-8c1c-4a9e-b7c4-8d4815a1568e","Type":"ContainerStarted","Data":"fddd47451bb81536addb9716c7934dfa314d3fdcbab8da4cc3500d25cc6bb2f6"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.754324 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.759911 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" podStartSLOduration=5.41739697 podStartE2EDuration="46.759885463s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.866968374 +0000 UTC m=+980.509169377" lastFinishedPulling="2025-12-08 19:38:59.209456867 +0000 UTC m=+1021.851657870" observedRunningTime="2025-12-08 19:39:01.746537232 +0000 UTC m=+1024.388738235" watchObservedRunningTime="2025-12-08 19:39:01.759885463 +0000 UTC m=+1024.402086456" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.770872 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" event={"ID":"53d61380-3a1e-45e5-a039-9d339bcdb9ea","Type":"ContainerStarted","Data":"958dc3c7967c1aa948d0f3c4c66ea029c931071eeeceea72c31277b52d1a2a76"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.771752 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.794643 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" event={"ID":"ec6df814-0549-46a9-a9f3-2f94083bda52","Type":"ContainerStarted","Data":"c068cda3ebacbf0bf7aa04a89753c428d0f8f432aa602a838a82534bff82bff2"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.795611 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.807358 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" podStartSLOduration=46.807337101 podStartE2EDuration="46.807337101s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:39:01.793370581 +0000 UTC m=+1024.435571584" watchObservedRunningTime="2025-12-08 19:39:01.807337101 +0000 UTC m=+1024.449538104" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.828427 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" event={"ID":"b02b307a-7d4c-4d20-b388-18f6fc6ec0ee","Type":"ContainerStarted","Data":"52ca38e24adf112c3ee72af553413688d852b22eb1049c949338b27702aa6fbe"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.848416 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" event={"ID":"1f3c9d68-2093-45a5-886e-f19c43b3e8e9","Type":"ContainerStarted","Data":"af9757cd90edb5e571fdc55cec9cd9f8ef4821b2d2341840e30d252d4aabc490"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.871642 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" event={"ID":"eedc4c77-93bc-40b9-ba3c-e7222d8828b5","Type":"ContainerStarted","Data":"81334cb5406054d6b01a761b07d12d230c10b03749f106cd3fae5f9ebd647cf6"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.884868 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" event={"ID":"beda813f-668e-45e8-8d1c-7c72caa6f294","Type":"ContainerStarted","Data":"7c168d55bf1da3686dc35659b829886db84c8dfdbb74e082922601f09c8d4898"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.886002 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.903817 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" event={"ID":"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3","Type":"ContainerStarted","Data":"7c7550576691386d80ec395855b6ec3ff965b010bf22761fefa3d5fe5bfd1d26"} Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.904552 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" podStartSLOduration=4.076838478 podStartE2EDuration="47.904521561s" podCreationTimestamp="2025-12-08 19:38:14 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.163093299 +0000 UTC m=+979.805294302" lastFinishedPulling="2025-12-08 19:39:00.990776382 +0000 UTC m=+1023.632977385" observedRunningTime="2025-12-08 19:39:01.893796244 +0000 UTC m=+1024.535997247" watchObservedRunningTime="2025-12-08 19:39:01.904521561 +0000 UTC m=+1024.546722564" Dec 08 19:39:01 crc kubenswrapper[4706]: E1208 19:39:01.908468 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:ccc60d56d8efc2e91a7d8a7131eb7e06c189c32247f2a819818c084ba2e2f2ab\\\"\"" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" podUID="ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3" Dec 08 19:39:01 crc kubenswrapper[4706]: E1208 19:39:01.908723 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" podUID="4fca7ad0-1e55-4b7a-bee5-08521a1836a2" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.977826 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" podStartSLOduration=5.22298988 podStartE2EDuration="46.977806687s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.461090144 +0000 UTC m=+980.103291137" lastFinishedPulling="2025-12-08 19:38:59.215906941 +0000 UTC m=+1021.858107944" observedRunningTime="2025-12-08 19:39:01.946031248 +0000 UTC m=+1024.588232251" watchObservedRunningTime="2025-12-08 19:39:01.977806687 +0000 UTC m=+1024.620007690" Dec 08 19:39:01 crc kubenswrapper[4706]: I1208 19:39:01.990731 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" podStartSLOduration=3.392214129 podStartE2EDuration="46.990710066s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.392364118 +0000 UTC m=+980.034565121" lastFinishedPulling="2025-12-08 19:39:00.990860055 +0000 UTC m=+1023.633061058" observedRunningTime="2025-12-08 19:39:01.977397746 +0000 UTC m=+1024.619598749" watchObservedRunningTime="2025-12-08 19:39:01.990710066 +0000 UTC m=+1024.632911069" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.015891 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" podStartSLOduration=5.683249786 podStartE2EDuration="47.015860596s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.875278772 +0000 UTC m=+980.517479775" lastFinishedPulling="2025-12-08 19:38:59.207889582 +0000 UTC m=+1021.850090585" observedRunningTime="2025-12-08 19:39:02.010654897 +0000 UTC m=+1024.652855900" watchObservedRunningTime="2025-12-08 19:39:02.015860596 +0000 UTC m=+1024.658061599" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.067650 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5xgcd" podStartSLOduration=31.400514408 podStartE2EDuration="47.067626657s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.411937718 +0000 UTC m=+980.054138721" lastFinishedPulling="2025-12-08 19:38:33.079049967 +0000 UTC m=+995.721250970" observedRunningTime="2025-12-08 19:39:02.063819748 +0000 UTC m=+1024.706020751" watchObservedRunningTime="2025-12-08 19:39:02.067626657 +0000 UTC m=+1024.709827680" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.120683 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-b7m9l" podStartSLOduration=32.261646315 podStartE2EDuration="48.120650494s" podCreationTimestamp="2025-12-08 19:38:14 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.205629876 +0000 UTC m=+979.847830879" lastFinishedPulling="2025-12-08 19:38:33.064634055 +0000 UTC m=+995.706835058" observedRunningTime="2025-12-08 19:39:02.113907571 +0000 UTC m=+1024.756108574" watchObservedRunningTime="2025-12-08 19:39:02.120650494 +0000 UTC m=+1024.762851497" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.195376 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5ndn8" podStartSLOduration=31.5352132 podStartE2EDuration="47.19535266s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.394469848 +0000 UTC m=+980.036670851" lastFinishedPulling="2025-12-08 19:38:33.054609308 +0000 UTC m=+995.696810311" observedRunningTime="2025-12-08 19:39:02.155195212 +0000 UTC m=+1024.797396215" watchObservedRunningTime="2025-12-08 19:39:02.19535266 +0000 UTC m=+1024.837553663" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.250997 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" podStartSLOduration=5.56768852 podStartE2EDuration="47.250973212s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.525674331 +0000 UTC m=+980.167875334" lastFinishedPulling="2025-12-08 19:38:59.208959023 +0000 UTC m=+1021.851160026" observedRunningTime="2025-12-08 19:39:02.199119688 +0000 UTC m=+1024.841320691" watchObservedRunningTime="2025-12-08 19:39:02.250973212 +0000 UTC m=+1024.893174225" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.912323 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" event={"ID":"1290df78-3e2c-494e-a264-c495d80c0bef","Type":"ContainerStarted","Data":"a947d2579a380bcb5fa139dd9f85c2f8661873675d2b31973b65c38778c7f71f"} Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.912985 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.916620 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" event={"ID":"8ddb889c-1432-4225-a110-e0aabc0ce3bc","Type":"ContainerStarted","Data":"3d43bc851725cd16d840022b1b60a7416fdde7069f96ea9482ab955df10ede23"} Dec 08 19:39:02 crc kubenswrapper[4706]: E1208 19:39:02.918558 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:ccc60d56d8efc2e91a7d8a7131eb7e06c189c32247f2a819818c084ba2e2f2ab\\\"\"" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" podUID="ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.948937 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" podStartSLOduration=5.315758648 podStartE2EDuration="48.948908657s" podCreationTimestamp="2025-12-08 19:38:14 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.459776556 +0000 UTC m=+980.101977559" lastFinishedPulling="2025-12-08 19:39:01.092926564 +0000 UTC m=+1023.735127568" observedRunningTime="2025-12-08 19:39:02.941041822 +0000 UTC m=+1025.583242825" watchObservedRunningTime="2025-12-08 19:39:02.948908657 +0000 UTC m=+1025.591109670" Dec 08 19:39:02 crc kubenswrapper[4706]: I1208 19:39:02.978252 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" podStartSLOduration=4.170635066 podStartE2EDuration="47.978230335s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.433868475 +0000 UTC m=+980.076069478" lastFinishedPulling="2025-12-08 19:39:01.241463744 +0000 UTC m=+1023.883664747" observedRunningTime="2025-12-08 19:39:02.976938059 +0000 UTC m=+1025.619139062" watchObservedRunningTime="2025-12-08 19:39:02.978230335 +0000 UTC m=+1025.620431338" Dec 08 19:39:03 crc kubenswrapper[4706]: I1208 19:39:03.924225 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" Dec 08 19:39:05 crc kubenswrapper[4706]: I1208 19:39:05.203428 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" Dec 08 19:39:05 crc kubenswrapper[4706]: I1208 19:39:05.206152 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" Dec 08 19:39:05 crc kubenswrapper[4706]: I1208 19:39:05.457147 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-pfq9b" Dec 08 19:39:05 crc kubenswrapper[4706]: I1208 19:39:05.488487 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kg2wq" Dec 08 19:39:05 crc kubenswrapper[4706]: I1208 19:39:05.836358 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:39:05 crc kubenswrapper[4706]: I1208 19:39:05.836423 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:39:05 crc kubenswrapper[4706]: I1208 19:39:05.941467 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" event={"ID":"4c42f8b9-a85c-4a45-a8f8-4a43e561a14e","Type":"ContainerStarted","Data":"f5c8b9df7277a9bac5bae2c01518b5debbf0cf39d7524241af02504ec0e809e2"} Dec 08 19:39:05 crc kubenswrapper[4706]: I1208 19:39:05.957177 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qtnlw" podStartSLOduration=35.544879295 podStartE2EDuration="51.95715663s" podCreationTimestamp="2025-12-08 19:38:14 +0000 UTC" firstStartedPulling="2025-12-08 19:38:16.643417944 +0000 UTC m=+979.285618947" lastFinishedPulling="2025-12-08 19:38:33.055695279 +0000 UTC m=+995.697896282" observedRunningTime="2025-12-08 19:39:05.956208902 +0000 UTC m=+1028.598409905" watchObservedRunningTime="2025-12-08 19:39:05.95715663 +0000 UTC m=+1028.599357633" Dec 08 19:39:06 crc kubenswrapper[4706]: I1208 19:39:06.050194 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vpwqp" Dec 08 19:39:06 crc kubenswrapper[4706]: I1208 19:39:06.311105 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5cbd74b459-5l5k7" Dec 08 19:39:06 crc kubenswrapper[4706]: I1208 19:39:06.564460 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qfsq9" Dec 08 19:39:06 crc kubenswrapper[4706]: I1208 19:39:06.572410 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" Dec 08 19:39:06 crc kubenswrapper[4706]: I1208 19:39:06.579901 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" Dec 08 19:39:06 crc kubenswrapper[4706]: I1208 19:39:06.956581 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" event={"ID":"ed8cb084-a66c-4fc0-afa1-068e58c80680","Type":"ContainerStarted","Data":"321aaf8b9e9d898637dabcb67c5b693d471e84ac3bba48107f0d2192e3f63722"} Dec 08 19:39:06 crc kubenswrapper[4706]: I1208 19:39:06.976507 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-g6gls" podStartSLOduration=36.752578526 podStartE2EDuration="51.976472877s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.851612005 +0000 UTC m=+980.493813018" lastFinishedPulling="2025-12-08 19:38:33.075506366 +0000 UTC m=+995.717707369" observedRunningTime="2025-12-08 19:39:06.972977997 +0000 UTC m=+1029.615179000" watchObservedRunningTime="2025-12-08 19:39:06.976472877 +0000 UTC m=+1029.618673900" Dec 08 19:39:08 crc kubenswrapper[4706]: I1208 19:39:08.100728 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-b89665ccc-lvggm" Dec 08 19:39:10 crc kubenswrapper[4706]: E1208 19:39:10.615036 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" podUID="9cca1f0f-9037-49d5-b8da-392421a67702" Dec 08 19:39:10 crc kubenswrapper[4706]: I1208 19:39:10.996678 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" event={"ID":"0f144ce9-7cc0-41f3-84ea-654579d32c0b","Type":"ContainerStarted","Data":"dc2b0d5272317fd10a96720f9a97719177cc1c58b6d6177ed4be51842d7a3747"} Dec 08 19:39:11 crc kubenswrapper[4706]: I1208 19:39:11.026051 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-7fvjs" podStartSLOduration=41.121191448 podStartE2EDuration="57.026003457s" podCreationTimestamp="2025-12-08 19:38:14 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.150383656 +0000 UTC m=+979.792584659" lastFinishedPulling="2025-12-08 19:38:33.055195665 +0000 UTC m=+995.697396668" observedRunningTime="2025-12-08 19:39:11.021739265 +0000 UTC m=+1033.663940268" watchObservedRunningTime="2025-12-08 19:39:11.026003457 +0000 UTC m=+1033.668204500" Dec 08 19:39:13 crc kubenswrapper[4706]: E1208 19:39:13.611139 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" podUID="6da8aab3-b971-413d-8557-cdfad2c95d2c" Dec 08 19:39:13 crc kubenswrapper[4706]: E1208 19:39:13.611158 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" podUID="22335991-2ba5-4fba-a1d1-b391e4767961" Dec 08 19:39:15 crc kubenswrapper[4706]: I1208 19:39:15.220815 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-j2mjl" Dec 08 19:39:15 crc kubenswrapper[4706]: I1208 19:39:15.457320 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cggng" Dec 08 19:39:15 crc kubenswrapper[4706]: I1208 19:39:15.781868 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-l8qxc" Dec 08 19:39:17 crc kubenswrapper[4706]: I1208 19:39:17.048667 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" event={"ID":"4fca7ad0-1e55-4b7a-bee5-08521a1836a2","Type":"ContainerStarted","Data":"cfd63d893aa8df9fc902e1f94d047a64965f4dc93757b258ecd270c47baaddad"} Dec 08 19:39:17 crc kubenswrapper[4706]: I1208 19:39:17.049820 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:39:17 crc kubenswrapper[4706]: I1208 19:39:17.086498 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" podStartSLOduration=20.115483072 podStartE2EDuration="1m2.086473742s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:34.039495332 +0000 UTC m=+996.681696325" lastFinishedPulling="2025-12-08 19:39:16.010485992 +0000 UTC m=+1038.652686995" observedRunningTime="2025-12-08 19:39:17.079036859 +0000 UTC m=+1039.721237892" watchObservedRunningTime="2025-12-08 19:39:17.086473742 +0000 UTC m=+1039.728674755" Dec 08 19:39:19 crc kubenswrapper[4706]: I1208 19:39:19.074086 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" event={"ID":"ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3","Type":"ContainerStarted","Data":"92b86b05759621a3b02dbc0ab39ae5129e08e49e1d9faf3e3f2fc132e974b82a"} Dec 08 19:39:19 crc kubenswrapper[4706]: I1208 19:39:19.075506 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:39:19 crc kubenswrapper[4706]: I1208 19:39:19.094965 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" podStartSLOduration=22.219031968 podStartE2EDuration="1m5.094946635s" podCreationTimestamp="2025-12-08 19:38:14 +0000 UTC" firstStartedPulling="2025-12-08 19:38:35.370243988 +0000 UTC m=+998.012444991" lastFinishedPulling="2025-12-08 19:39:18.246158645 +0000 UTC m=+1040.888359658" observedRunningTime="2025-12-08 19:39:19.093475123 +0000 UTC m=+1041.735676156" watchObservedRunningTime="2025-12-08 19:39:19.094946635 +0000 UTC m=+1041.737147638" Dec 08 19:39:21 crc kubenswrapper[4706]: I1208 19:39:21.914951 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f6nqtx" Dec 08 19:39:26 crc kubenswrapper[4706]: I1208 19:39:26.132236 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" event={"ID":"22335991-2ba5-4fba-a1d1-b391e4767961","Type":"ContainerStarted","Data":"b12a9eb5d27f749eb66ccb7a0c1bf4fe1e3879b8f269ba7bbe93ff96a00516b2"} Dec 08 19:39:26 crc kubenswrapper[4706]: I1208 19:39:26.133289 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" Dec 08 19:39:26 crc kubenswrapper[4706]: I1208 19:39:26.154374 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" podStartSLOduration=3.791909917 podStartE2EDuration="1m11.154353049s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.732895699 +0000 UTC m=+980.375096702" lastFinishedPulling="2025-12-08 19:39:25.095338821 +0000 UTC m=+1047.737539834" observedRunningTime="2025-12-08 19:39:26.152577069 +0000 UTC m=+1048.794778072" watchObservedRunningTime="2025-12-08 19:39:26.154353049 +0000 UTC m=+1048.796554052" Dec 08 19:39:27 crc kubenswrapper[4706]: I1208 19:39:27.144191 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" event={"ID":"9cca1f0f-9037-49d5-b8da-392421a67702","Type":"ContainerStarted","Data":"e93348f0c69dc7a404a1e23105358ea0f7016f2635c4f3cbd692a498d7bbdfe9"} Dec 08 19:39:27 crc kubenswrapper[4706]: I1208 19:39:27.175572 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bqsg2" podStartSLOduration=4.029480561 podStartE2EDuration="1m12.175544141s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.887059879 +0000 UTC m=+980.529260882" lastFinishedPulling="2025-12-08 19:39:26.033123459 +0000 UTC m=+1048.675324462" observedRunningTime="2025-12-08 19:39:27.166985897 +0000 UTC m=+1049.809186910" watchObservedRunningTime="2025-12-08 19:39:27.175544141 +0000 UTC m=+1049.817745154" Dec 08 19:39:30 crc kubenswrapper[4706]: I1208 19:39:30.975725 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-qmcdh" Dec 08 19:39:32 crc kubenswrapper[4706]: I1208 19:39:32.203000 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" event={"ID":"6da8aab3-b971-413d-8557-cdfad2c95d2c","Type":"ContainerStarted","Data":"e4eb61ec94cbb3b0aa3c42bf5d5184b889bd360d279cad8da4ed8c5d94492151"} Dec 08 19:39:32 crc kubenswrapper[4706]: I1208 19:39:32.203683 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" Dec 08 19:39:32 crc kubenswrapper[4706]: I1208 19:39:32.235791 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" podStartSLOduration=3.558356951 podStartE2EDuration="1m17.235761597s" podCreationTimestamp="2025-12-08 19:38:15 +0000 UTC" firstStartedPulling="2025-12-08 19:38:17.737142691 +0000 UTC m=+980.379343694" lastFinishedPulling="2025-12-08 19:39:31.414547337 +0000 UTC m=+1054.056748340" observedRunningTime="2025-12-08 19:39:32.22744084 +0000 UTC m=+1054.869641853" watchObservedRunningTime="2025-12-08 19:39:32.235761597 +0000 UTC m=+1054.877962600" Dec 08 19:39:35 crc kubenswrapper[4706]: I1208 19:39:35.836689 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:39:35 crc kubenswrapper[4706]: I1208 19:39:35.837494 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:39:35 crc kubenswrapper[4706]: I1208 19:39:35.837562 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:39:35 crc kubenswrapper[4706]: I1208 19:39:35.838445 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"44273cc3a85c65641d37bbdee9218aa6e6f7747d6e135429b54fe362e247061f"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:39:35 crc kubenswrapper[4706]: I1208 19:39:35.838521 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://44273cc3a85c65641d37bbdee9218aa6e6f7747d6e135429b54fe362e247061f" gracePeriod=600 Dec 08 19:39:36 crc kubenswrapper[4706]: I1208 19:39:36.066210 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-58rkn" Dec 08 19:39:36 crc kubenswrapper[4706]: I1208 19:39:36.248968 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="44273cc3a85c65641d37bbdee9218aa6e6f7747d6e135429b54fe362e247061f" exitCode=0 Dec 08 19:39:36 crc kubenswrapper[4706]: I1208 19:39:36.249028 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"44273cc3a85c65641d37bbdee9218aa6e6f7747d6e135429b54fe362e247061f"} Dec 08 19:39:36 crc kubenswrapper[4706]: I1208 19:39:36.249064 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"5cdb7c8ca56123f7a237303ad544e086627575cfd4a9196d909f15ed3eb4ca79"} Dec 08 19:39:36 crc kubenswrapper[4706]: I1208 19:39:36.249087 4706 scope.go:117] "RemoveContainer" containerID="7d44a6096d27dcbd84b0ab9d9f054b00bc8d565b62f431d2f781917837a86f52" Dec 08 19:39:36 crc kubenswrapper[4706]: I1208 19:39:36.333402 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nbcmr" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.432788 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n2vgp"] Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.435370 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.438202 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.438532 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-p82w5" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.440905 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.441211 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.445680 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n2vgp"] Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.513491 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q5fhn"] Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.515438 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.519343 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.533996 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q5fhn"] Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.534316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acfc18e7-14e4-4aff-b304-b293233bb76e-config\") pod \"dnsmasq-dns-675f4bcbfc-n2vgp\" (UID: \"acfc18e7-14e4-4aff-b304-b293233bb76e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.534485 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp7jc\" (UniqueName: \"kubernetes.io/projected/acfc18e7-14e4-4aff-b304-b293233bb76e-kube-api-access-dp7jc\") pod \"dnsmasq-dns-675f4bcbfc-n2vgp\" (UID: \"acfc18e7-14e4-4aff-b304-b293233bb76e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.636070 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.636148 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5fwh\" (UniqueName: \"kubernetes.io/projected/dd4d903d-5735-40d1-9131-e587537401cf-kube-api-access-x5fwh\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.636193 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp7jc\" (UniqueName: \"kubernetes.io/projected/acfc18e7-14e4-4aff-b304-b293233bb76e-kube-api-access-dp7jc\") pod \"dnsmasq-dns-675f4bcbfc-n2vgp\" (UID: \"acfc18e7-14e4-4aff-b304-b293233bb76e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.636244 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acfc18e7-14e4-4aff-b304-b293233bb76e-config\") pod \"dnsmasq-dns-675f4bcbfc-n2vgp\" (UID: \"acfc18e7-14e4-4aff-b304-b293233bb76e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.636287 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-config\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.637372 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acfc18e7-14e4-4aff-b304-b293233bb76e-config\") pod \"dnsmasq-dns-675f4bcbfc-n2vgp\" (UID: \"acfc18e7-14e4-4aff-b304-b293233bb76e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.688712 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp7jc\" (UniqueName: \"kubernetes.io/projected/acfc18e7-14e4-4aff-b304-b293233bb76e-kube-api-access-dp7jc\") pod \"dnsmasq-dns-675f4bcbfc-n2vgp\" (UID: \"acfc18e7-14e4-4aff-b304-b293233bb76e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.738186 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-config\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.738288 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.738355 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5fwh\" (UniqueName: \"kubernetes.io/projected/dd4d903d-5735-40d1-9131-e587537401cf-kube-api-access-x5fwh\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.739255 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-config\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.739283 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.755529 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.757710 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5fwh\" (UniqueName: \"kubernetes.io/projected/dd4d903d-5735-40d1-9131-e587537401cf-kube-api-access-x5fwh\") pod \"dnsmasq-dns-78dd6ddcc-q5fhn\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:50 crc kubenswrapper[4706]: I1208 19:39:50.835820 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:39:51 crc kubenswrapper[4706]: I1208 19:39:51.379511 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n2vgp"] Dec 08 19:39:51 crc kubenswrapper[4706]: I1208 19:39:51.527195 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q5fhn"] Dec 08 19:39:51 crc kubenswrapper[4706]: W1208 19:39:51.535175 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd4d903d_5735_40d1_9131_e587537401cf.slice/crio-4d92f7fd0d0660140c731a5108b834158909273407288d582cec2fb1730e857e WatchSource:0}: Error finding container 4d92f7fd0d0660140c731a5108b834158909273407288d582cec2fb1730e857e: Status 404 returned error can't find the container with id 4d92f7fd0d0660140c731a5108b834158909273407288d582cec2fb1730e857e Dec 08 19:39:52 crc kubenswrapper[4706]: I1208 19:39:52.394028 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" event={"ID":"acfc18e7-14e4-4aff-b304-b293233bb76e","Type":"ContainerStarted","Data":"9a58bc443407e1958b49f1aa31c2a7ac05dc6a8971cf29f91f003703f1672d70"} Dec 08 19:39:52 crc kubenswrapper[4706]: I1208 19:39:52.395905 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" event={"ID":"dd4d903d-5735-40d1-9131-e587537401cf","Type":"ContainerStarted","Data":"4d92f7fd0d0660140c731a5108b834158909273407288d582cec2fb1730e857e"} Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.606313 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n2vgp"] Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.638373 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9kbjh"] Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.642976 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.648357 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9kbjh"] Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.805658 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-config\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.806162 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.806208 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmgtv\" (UniqueName: \"kubernetes.io/projected/b9f26f6c-226a-4026-bde7-10f4401f53e0-kube-api-access-nmgtv\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.911847 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.911926 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmgtv\" (UniqueName: \"kubernetes.io/projected/b9f26f6c-226a-4026-bde7-10f4401f53e0-kube-api-access-nmgtv\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.912024 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-config\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.915310 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.917793 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-config\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.968459 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmgtv\" (UniqueName: \"kubernetes.io/projected/b9f26f6c-226a-4026-bde7-10f4401f53e0-kube-api-access-nmgtv\") pod \"dnsmasq-dns-666b6646f7-9kbjh\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:53 crc kubenswrapper[4706]: I1208 19:39:53.972694 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.088368 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q5fhn"] Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.140941 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lqbnv"] Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.143367 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.211675 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lqbnv"] Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.228112 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-config\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.231423 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.231603 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz9j6\" (UniqueName: \"kubernetes.io/projected/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-kube-api-access-qz9j6\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.333972 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-config\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.334578 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.334621 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz9j6\" (UniqueName: \"kubernetes.io/projected/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-kube-api-access-qz9j6\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.335375 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-config\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.336135 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.412981 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz9j6\" (UniqueName: \"kubernetes.io/projected/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-kube-api-access-qz9j6\") pod \"dnsmasq-dns-57d769cc4f-lqbnv\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.477248 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.729003 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9kbjh"] Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.835233 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.839869 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.844156 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.844497 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.844668 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.844812 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.844939 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-cl2tg" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.845074 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.845378 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.849859 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.945496 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.945573 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1472b8cb-795a-4a57-95f8-184a7707e8ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.945601 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.945630 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.945666 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvb7f\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-kube-api-access-hvb7f\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.945699 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.946000 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.946107 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.946129 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.946178 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1472b8cb-795a-4a57-95f8-184a7707e8ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:54 crc kubenswrapper[4706]: I1208 19:39:54.946374 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.049211 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lqbnv"] Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.050848 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1472b8cb-795a-4a57-95f8-184a7707e8ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.050937 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.050994 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.051080 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvb7f\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-kube-api-access-hvb7f\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.051114 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.051170 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.051214 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.051248 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.051299 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1472b8cb-795a-4a57-95f8-184a7707e8ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.051340 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.051412 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.060610 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: W1208 19:39:55.061444 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52a95e2d_d2d3_4a2e_9e4a_bd3238ed9568.slice/crio-45e0293d8502ead0e496e9839adaca9aaaea39509c8f6860e7935e13cc20f8f4 WatchSource:0}: Error finding container 45e0293d8502ead0e496e9839adaca9aaaea39509c8f6860e7935e13cc20f8f4: Status 404 returned error can't find the container with id 45e0293d8502ead0e496e9839adaca9aaaea39509c8f6860e7935e13cc20f8f4 Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.063840 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.065005 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.069557 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.069608 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0cccd51a6a12e5e48b5c2f1604e3858ae97fd79f1ff6b61d3ff2b3945f701dee/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.070127 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1472b8cb-795a-4a57-95f8-184a7707e8ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.079729 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.080542 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.081042 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1472b8cb-795a-4a57-95f8-184a7707e8ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.081898 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.082329 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.089553 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvb7f\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-kube-api-access-hvb7f\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.154216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"rabbitmq-server-0\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.186419 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.332084 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.334544 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.339690 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.347100 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.347552 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-glhtm" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.347824 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.347972 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.348428 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.348739 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.358850 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464753 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ea8337d-cff3-414c-98f6-17b41653b4fb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464812 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464836 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ea8337d-cff3-414c-98f6-17b41653b4fb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464859 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464884 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464907 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464926 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464948 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.464970 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhlbw\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-kube-api-access-rhlbw\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.465007 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.465048 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.467490 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" event={"ID":"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568","Type":"ContainerStarted","Data":"45e0293d8502ead0e496e9839adaca9aaaea39509c8f6860e7935e13cc20f8f4"} Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.470380 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" event={"ID":"b9f26f6c-226a-4026-bde7-10f4401f53e0","Type":"ContainerStarted","Data":"510b78f609ebee0e06c1af21a7d7aebc9d74783f27b4e0f40f2137e8c575b1f5"} Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567018 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567179 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ea8337d-cff3-414c-98f6-17b41653b4fb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567220 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567251 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ea8337d-cff3-414c-98f6-17b41653b4fb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567328 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567366 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567418 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567446 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567471 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567503 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhlbw\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-kube-api-access-rhlbw\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.567532 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.570112 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.570351 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.570442 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.570440 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.571173 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.578117 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.578173 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ea8337d-cff3-414c-98f6-17b41653b4fb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.578188 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.578319 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8f2cec159fcbaca32539204793034491d123d827594a614a5fcc96cb10f96558/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.578575 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ea8337d-cff3-414c-98f6-17b41653b4fb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.578940 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.589754 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhlbw\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-kube-api-access-rhlbw\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.625423 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.679253 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:39:55 crc kubenswrapper[4706]: I1208 19:39:55.847667 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:39:55 crc kubenswrapper[4706]: W1208 19:39:55.961554 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1472b8cb_795a_4a57_95f8_184a7707e8ad.slice/crio-d4e9fc451def0cba335a61435bed09c8b9d26cd0d42bfb2e536fff764f9ef677 WatchSource:0}: Error finding container d4e9fc451def0cba335a61435bed09c8b9d26cd0d42bfb2e536fff764f9ef677: Status 404 returned error can't find the container with id d4e9fc451def0cba335a61435bed09c8b9d26cd0d42bfb2e536fff764f9ef677 Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.259329 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.360526 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.363359 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.403082 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-ljrps" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.408297 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.408704 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.408955 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.411628 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.414179 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.499719 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1472b8cb-795a-4a57-95f8-184a7707e8ad","Type":"ContainerStarted","Data":"d4e9fc451def0cba335a61435bed09c8b9d26cd0d42bfb2e536fff764f9ef677"} Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.505726 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.505788 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fa44e847-3503-421e-9810-2b746480f0c3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fa44e847-3503-421e-9810-2b746480f0c3\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.505825 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.505864 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.505899 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.506439 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt4p2\" (UniqueName: \"kubernetes.io/projected/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-kube-api-access-vt4p2\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.506475 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.506532 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.607765 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.607825 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fa44e847-3503-421e-9810-2b746480f0c3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fa44e847-3503-421e-9810-2b746480f0c3\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.607879 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.607910 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.608185 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.608228 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt4p2\" (UniqueName: \"kubernetes.io/projected/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-kube-api-access-vt4p2\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.608246 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.608302 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.616672 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.617133 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.617758 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.617899 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.625806 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.627245 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.628076 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fa44e847-3503-421e-9810-2b746480f0c3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fa44e847-3503-421e-9810-2b746480f0c3\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0ee780f00918f90f1d21d3e57b9f6062315601f2bdb99b2e4e46f40837a86c97/globalmount\"" pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.634879 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt4p2\" (UniqueName: \"kubernetes.io/projected/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-kube-api-access-vt4p2\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.635422 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/156710b7-7dbe-4bb7-9c69-ce726d4b7a4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.673526 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fa44e847-3503-421e-9810-2b746480f0c3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fa44e847-3503-421e-9810-2b746480f0c3\") pod \"openstack-galera-0\" (UID: \"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c\") " pod="openstack/openstack-galera-0" Dec 08 19:39:56 crc kubenswrapper[4706]: I1208 19:39:56.739451 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.913859 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.915909 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.928124 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8cxh2" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.928570 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.928698 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.928815 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.938520 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13532f07-b125-4d67-bdb8-006b583bea40-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.938570 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nqxv\" (UniqueName: \"kubernetes.io/projected/13532f07-b125-4d67-bdb8-006b583bea40-kube-api-access-2nqxv\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.938595 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/13532f07-b125-4d67-bdb8-006b583bea40-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.938622 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.938938 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/13532f07-b125-4d67-bdb8-006b583bea40-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.939047 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.939115 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3d1e6a6d-e451-4505-bc0d-bbcc2213d928\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3d1e6a6d-e451-4505-bc0d-bbcc2213d928\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.939169 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:57 crc kubenswrapper[4706]: I1208 19:39:57.943389 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.041236 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/13532f07-b125-4d67-bdb8-006b583bea40-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.041306 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.041342 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3d1e6a6d-e451-4505-bc0d-bbcc2213d928\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3d1e6a6d-e451-4505-bc0d-bbcc2213d928\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.041369 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.041407 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13532f07-b125-4d67-bdb8-006b583bea40-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.041430 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nqxv\" (UniqueName: \"kubernetes.io/projected/13532f07-b125-4d67-bdb8-006b583bea40-kube-api-access-2nqxv\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.041452 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/13532f07-b125-4d67-bdb8-006b583bea40-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.041477 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.042627 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/13532f07-b125-4d67-bdb8-006b583bea40-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.043858 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.044322 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.045923 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13532f07-b125-4d67-bdb8-006b583bea40-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.046477 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13532f07-b125-4d67-bdb8-006b583bea40-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.046884 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/13532f07-b125-4d67-bdb8-006b583bea40-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.049177 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.049219 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3d1e6a6d-e451-4505-bc0d-bbcc2213d928\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3d1e6a6d-e451-4505-bc0d-bbcc2213d928\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/98fc4c497ee4a26ec69e6da46193b781c1c5b1c16bbe70d4c54da61e9e060202/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.090995 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nqxv\" (UniqueName: \"kubernetes.io/projected/13532f07-b125-4d67-bdb8-006b583bea40-kube-api-access-2nqxv\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.121296 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3d1e6a6d-e451-4505-bc0d-bbcc2213d928\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3d1e6a6d-e451-4505-bc0d-bbcc2213d928\") pod \"openstack-cell1-galera-0\" (UID: \"13532f07-b125-4d67-bdb8-006b583bea40\") " pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.143207 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.145679 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.151056 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.151340 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-tvqkm" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.151438 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.155989 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.244552 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fj6k\" (UniqueName: \"kubernetes.io/projected/fd927fec-2b54-41d2-affb-19f9fd8c96ab-kube-api-access-7fj6k\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.244683 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd927fec-2b54-41d2-affb-19f9fd8c96ab-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.244750 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd927fec-2b54-41d2-affb-19f9fd8c96ab-config-data\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.244777 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd927fec-2b54-41d2-affb-19f9fd8c96ab-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.244806 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd927fec-2b54-41d2-affb-19f9fd8c96ab-kolla-config\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.258196 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.347420 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fj6k\" (UniqueName: \"kubernetes.io/projected/fd927fec-2b54-41d2-affb-19f9fd8c96ab-kube-api-access-7fj6k\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.347548 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd927fec-2b54-41d2-affb-19f9fd8c96ab-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.347598 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd927fec-2b54-41d2-affb-19f9fd8c96ab-config-data\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.347627 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd927fec-2b54-41d2-affb-19f9fd8c96ab-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.347647 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd927fec-2b54-41d2-affb-19f9fd8c96ab-kolla-config\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.348452 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd927fec-2b54-41d2-affb-19f9fd8c96ab-kolla-config\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.349197 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd927fec-2b54-41d2-affb-19f9fd8c96ab-config-data\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.363824 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd927fec-2b54-41d2-affb-19f9fd8c96ab-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.371677 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd927fec-2b54-41d2-affb-19f9fd8c96ab-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.377229 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fj6k\" (UniqueName: \"kubernetes.io/projected/fd927fec-2b54-41d2-affb-19f9fd8c96ab-kube-api-access-7fj6k\") pod \"memcached-0\" (UID: \"fd927fec-2b54-41d2-affb-19f9fd8c96ab\") " pod="openstack/memcached-0" Dec 08 19:39:58 crc kubenswrapper[4706]: I1208 19:39:58.491977 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 08 19:40:00 crc kubenswrapper[4706]: I1208 19:40:00.322677 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:40:00 crc kubenswrapper[4706]: I1208 19:40:00.324176 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 19:40:00 crc kubenswrapper[4706]: I1208 19:40:00.327473 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-kndvz" Dec 08 19:40:00 crc kubenswrapper[4706]: I1208 19:40:00.337349 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:40:00 crc kubenswrapper[4706]: I1208 19:40:00.387231 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s94p\" (UniqueName: \"kubernetes.io/projected/908bed13-e8ef-4234-83fd-21b3f0ed218e-kube-api-access-8s94p\") pod \"kube-state-metrics-0\" (UID: \"908bed13-e8ef-4234-83fd-21b3f0ed218e\") " pod="openstack/kube-state-metrics-0" Dec 08 19:40:00 crc kubenswrapper[4706]: I1208 19:40:00.490433 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s94p\" (UniqueName: \"kubernetes.io/projected/908bed13-e8ef-4234-83fd-21b3f0ed218e-kube-api-access-8s94p\") pod \"kube-state-metrics-0\" (UID: \"908bed13-e8ef-4234-83fd-21b3f0ed218e\") " pod="openstack/kube-state-metrics-0" Dec 08 19:40:00 crc kubenswrapper[4706]: I1208 19:40:00.545173 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s94p\" (UniqueName: \"kubernetes.io/projected/908bed13-e8ef-4234-83fd-21b3f0ed218e-kube-api-access-8s94p\") pod \"kube-state-metrics-0\" (UID: \"908bed13-e8ef-4234-83fd-21b3f0ed218e\") " pod="openstack/kube-state-metrics-0" Dec 08 19:40:00 crc kubenswrapper[4706]: I1208 19:40:00.646553 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.178666 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.183519 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.200868 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.201062 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.201346 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.201408 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.201589 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-l9qrj" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.212037 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.330475 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88233d80-672d-47c3-9a26-70c22f748e29-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.330543 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/88233d80-672d-47c3-9a26-70c22f748e29-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.330636 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88233d80-672d-47c3-9a26-70c22f748e29-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.330664 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.330708 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbq8t\" (UniqueName: \"kubernetes.io/projected/88233d80-672d-47c3-9a26-70c22f748e29-kube-api-access-mbq8t\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.330735 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.330864 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.433053 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88233d80-672d-47c3-9a26-70c22f748e29-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.433132 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/88233d80-672d-47c3-9a26-70c22f748e29-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.433242 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88233d80-672d-47c3-9a26-70c22f748e29-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.433310 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.433360 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbq8t\" (UniqueName: \"kubernetes.io/projected/88233d80-672d-47c3-9a26-70c22f748e29-kube-api-access-mbq8t\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.433384 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.433443 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.437038 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/88233d80-672d-47c3-9a26-70c22f748e29-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.441211 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.441722 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88233d80-672d-47c3-9a26-70c22f748e29-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.442157 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.442977 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88233d80-672d-47c3-9a26-70c22f748e29-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.443475 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/88233d80-672d-47c3-9a26-70c22f748e29-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.464831 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbq8t\" (UniqueName: \"kubernetes.io/projected/88233d80-672d-47c3-9a26-70c22f748e29-kube-api-access-mbq8t\") pod \"alertmanager-metric-storage-0\" (UID: \"88233d80-672d-47c3-9a26-70c22f748e29\") " pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.518647 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.725679 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.728570 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.732325 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.732496 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-8tk42" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.732611 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.732790 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.736459 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.751110 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.753201 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.843230 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.843377 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.843420 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4853c157-69b5-45ec-8fad-9e5ddea58cea-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.843485 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-config\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.843713 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp9jl\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-kube-api-access-zp9jl\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.843815 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.843950 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4853c157-69b5-45ec-8fad-9e5ddea58cea-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.843987 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.946182 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.946251 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.946301 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4853c157-69b5-45ec-8fad-9e5ddea58cea-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.946349 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-config\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.946401 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp9jl\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-kube-api-access-zp9jl\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.946425 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.946467 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4853c157-69b5-45ec-8fad-9e5ddea58cea-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.946490 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.948102 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4853c157-69b5-45ec-8fad-9e5ddea58cea-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.950362 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.950743 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e53f9d22d9359bccec8cb0090060887d02148bfeb3788e3e61a1431ce860cbd0/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.953447 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4853c157-69b5-45ec-8fad-9e5ddea58cea-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.955947 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.959573 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.959770 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.972721 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-config\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:01 crc kubenswrapper[4706]: I1208 19:40:01.973832 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp9jl\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-kube-api-access-zp9jl\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:02 crc kubenswrapper[4706]: I1208 19:40:02.016897 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"prometheus-metric-storage-0\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:02 crc kubenswrapper[4706]: I1208 19:40:02.059781 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.075569 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-78wps"] Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.086866 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.095161 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-vwf4b" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.095423 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.095570 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.113570 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78wps"] Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.126289 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-log-ovn\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.126357 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-run-ovn\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.126411 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79f551cd-d08d-4739-9041-4543f05fbb60-scripts\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.126435 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7c2m\" (UniqueName: \"kubernetes.io/projected/79f551cd-d08d-4739-9041-4543f05fbb60-kube-api-access-l7c2m\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.126483 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-run\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.126513 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79f551cd-d08d-4739-9041-4543f05fbb60-ovn-controller-tls-certs\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.126550 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f551cd-d08d-4739-9041-4543f05fbb60-combined-ca-bundle\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.173341 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.175175 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.189543 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.190024 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-dp27x" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.190102 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.190247 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.190357 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.200386 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.228581 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-run\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.228660 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79f551cd-d08d-4739-9041-4543f05fbb60-ovn-controller-tls-certs\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.228735 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f551cd-d08d-4739-9041-4543f05fbb60-combined-ca-bundle\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.228835 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-log-ovn\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.228930 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-run-ovn\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.229000 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79f551cd-d08d-4739-9041-4543f05fbb60-scripts\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.229022 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7c2m\" (UniqueName: \"kubernetes.io/projected/79f551cd-d08d-4739-9041-4543f05fbb60-kube-api-access-l7c2m\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.229960 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-run\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.230073 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-run-ovn\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.230386 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79f551cd-d08d-4739-9041-4543f05fbb60-var-log-ovn\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.232111 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79f551cd-d08d-4739-9041-4543f05fbb60-scripts\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.242478 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79f551cd-d08d-4739-9041-4543f05fbb60-ovn-controller-tls-certs\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.255578 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f551cd-d08d-4739-9041-4543f05fbb60-combined-ca-bundle\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.266054 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7c2m\" (UniqueName: \"kubernetes.io/projected/79f551cd-d08d-4739-9041-4543f05fbb60-kube-api-access-l7c2m\") pod \"ovn-controller-78wps\" (UID: \"79f551cd-d08d-4739-9041-4543f05fbb60\") " pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.319247 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-fb6gx"] Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.322233 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.330430 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.330538 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.330563 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f498bda7-c379-4808-bd6c-8e5513c4322d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.330598 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhbfx\" (UniqueName: \"kubernetes.io/projected/f498bda7-c379-4808-bd6c-8e5513c4322d-kube-api-access-dhbfx\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.330629 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9ed1a02f-4f57-4620-99db-d4fa6c9c3656\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ed1a02f-4f57-4620-99db-d4fa6c9c3656\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.330658 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f498bda7-c379-4808-bd6c-8e5513c4322d-config\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.330678 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f498bda7-c379-4808-bd6c-8e5513c4322d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.330736 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.347674 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fb6gx"] Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.417004 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78wps" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.432747 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-etc-ovs\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.432811 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-run\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.432849 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhbfx\" (UniqueName: \"kubernetes.io/projected/f498bda7-c379-4808-bd6c-8e5513c4322d-kube-api-access-dhbfx\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.432883 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9ed1a02f-4f57-4620-99db-d4fa6c9c3656\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ed1a02f-4f57-4620-99db-d4fa6c9c3656\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.432909 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f498bda7-c379-4808-bd6c-8e5513c4322d-config\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.432929 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f498bda7-c379-4808-bd6c-8e5513c4322d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.432952 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-log\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.432988 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25287a9f-27c5-47a7-945e-684eea0157c8-scripts\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.433027 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.433052 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.433087 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-lib\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.433106 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbsc4\" (UniqueName: \"kubernetes.io/projected/25287a9f-27c5-47a7-945e-684eea0157c8-kube-api-access-xbsc4\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.433126 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.433142 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f498bda7-c379-4808-bd6c-8e5513c4322d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.434412 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f498bda7-c379-4808-bd6c-8e5513c4322d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.435455 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f498bda7-c379-4808-bd6c-8e5513c4322d-config\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.436108 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f498bda7-c379-4808-bd6c-8e5513c4322d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.470006 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.480545 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhbfx\" (UniqueName: \"kubernetes.io/projected/f498bda7-c379-4808-bd6c-8e5513c4322d-kube-api-access-dhbfx\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.495318 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.502226 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498bda7-c379-4808-bd6c-8e5513c4322d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.534599 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25287a9f-27c5-47a7-945e-684eea0157c8-scripts\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.534736 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-lib\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.534761 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbsc4\" (UniqueName: \"kubernetes.io/projected/25287a9f-27c5-47a7-945e-684eea0157c8-kube-api-access-xbsc4\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.534811 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-etc-ovs\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.534840 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-run\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.534910 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-log\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.535446 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-log\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.537429 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25287a9f-27c5-47a7-945e-684eea0157c8-scripts\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.537575 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-lib\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.538029 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-etc-ovs\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.538070 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/25287a9f-27c5-47a7-945e-684eea0157c8-var-run\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.538864 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.538901 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9ed1a02f-4f57-4620-99db-d4fa6c9c3656\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ed1a02f-4f57-4620-99db-d4fa6c9c3656\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf2d6c83af5cc8d8b5a87507ec9ef15e7d591768c23b79b78e187c45e2fd38e8/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.562982 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbsc4\" (UniqueName: \"kubernetes.io/projected/25287a9f-27c5-47a7-945e-684eea0157c8-kube-api-access-xbsc4\") pod \"ovn-controller-ovs-fb6gx\" (UID: \"25287a9f-27c5-47a7-945e-684eea0157c8\") " pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.597126 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9ed1a02f-4f57-4620-99db-d4fa6c9c3656\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ed1a02f-4f57-4620-99db-d4fa6c9c3656\") pod \"ovsdbserver-nb-0\" (UID: \"f498bda7-c379-4808-bd6c-8e5513c4322d\") " pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.656091 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:04 crc kubenswrapper[4706]: I1208 19:40:04.801489 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:08 crc kubenswrapper[4706]: I1208 19:40:08.952342 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 08 19:40:08 crc kubenswrapper[4706]: I1208 19:40:08.954761 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:08 crc kubenswrapper[4706]: I1208 19:40:08.957315 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 08 19:40:08 crc kubenswrapper[4706]: I1208 19:40:08.957361 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 08 19:40:08 crc kubenswrapper[4706]: I1208 19:40:08.959553 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 08 19:40:08 crc kubenswrapper[4706]: I1208 19:40:08.966357 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-n4svt" Dec 08 19:40:08 crc kubenswrapper[4706]: I1208 19:40:08.969751 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.016699 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-config\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.017105 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-72a246d1-e4d7-4a82-bc44-3d89988c9947\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-72a246d1-e4d7-4a82-bc44-3d89988c9947\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.017328 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.017462 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cvx4\" (UniqueName: \"kubernetes.io/projected/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-kube-api-access-5cvx4\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.017571 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.017733 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.017853 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.017982 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.120283 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-config\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.120356 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-72a246d1-e4d7-4a82-bc44-3d89988c9947\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-72a246d1-e4d7-4a82-bc44-3d89988c9947\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.120401 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.120435 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cvx4\" (UniqueName: \"kubernetes.io/projected/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-kube-api-access-5cvx4\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.120456 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.120508 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.120540 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.120580 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.121431 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-config\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.122490 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.123140 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.130885 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.131303 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.131385 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.131411 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-72a246d1-e4d7-4a82-bc44-3d89988c9947\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-72a246d1-e4d7-4a82-bc44-3d89988c9947\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/14c3fe4058a363e4c3952b955047744016471508060529c227d0d9c4f5ff6a9e/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.135985 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.161238 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cvx4\" (UniqueName: \"kubernetes.io/projected/40f5d973-ac12-469a-a1d9-aa871b7cc9a3-kube-api-access-5cvx4\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.171806 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-72a246d1-e4d7-4a82-bc44-3d89988c9947\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-72a246d1-e4d7-4a82-bc44-3d89988c9947\") pod \"ovsdbserver-sb-0\" (UID: \"40f5d973-ac12-469a-a1d9-aa871b7cc9a3\") " pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: I1208 19:40:09.277160 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 08 19:40:09 crc kubenswrapper[4706]: W1208 19:40:09.815215 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ea8337d_cff3_414c_98f6_17b41653b4fb.slice/crio-a615604c3637a30de2798fafe858b135087d5f0699a97e6ac7ed4daf394adbd4 WatchSource:0}: Error finding container a615604c3637a30de2798fafe858b135087d5f0699a97e6ac7ed4daf394adbd4: Status 404 returned error can't find the container with id a615604c3637a30de2798fafe858b135087d5f0699a97e6ac7ed4daf394adbd4 Dec 08 19:40:10 crc kubenswrapper[4706]: I1208 19:40:10.697883 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ea8337d-cff3-414c-98f6-17b41653b4fb","Type":"ContainerStarted","Data":"a615604c3637a30de2798fafe858b135087d5f0699a97e6ac7ed4daf394adbd4"} Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.310494 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5"] Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.313905 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.317438 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-grpc" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.326655 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-dockercfg-sp4lz" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.327189 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca-bundle" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.327507 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-http" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.327808 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-config" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.346382 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5"] Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.397625 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.397683 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.397722 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf52g\" (UniqueName: \"kubernetes.io/projected/081f2271-cba5-43e4-a179-7fd4389c6531-kube-api-access-zf52g\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.397754 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.398182 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/081f2271-cba5-43e4-a179-7fd4389c6531-config\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.500423 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/081f2271-cba5-43e4-a179-7fd4389c6531-config\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.500515 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.500550 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.500587 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf52g\" (UniqueName: \"kubernetes.io/projected/081f2271-cba5-43e4-a179-7fd4389c6531-kube-api-access-zf52g\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.500624 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.502746 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/081f2271-cba5-43e4-a179-7fd4389c6531-config\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.508045 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.511820 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.515687 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/081f2271-cba5-43e4-a179-7fd4389c6531-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.567388 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm"] Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.568759 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.570859 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf52g\" (UniqueName: \"kubernetes.io/projected/081f2271-cba5-43e4-a179-7fd4389c6531-kube-api-access-zf52g\") pod \"cloudkitty-lokistack-distributor-664b687b54-zfbc5\" (UID: \"081f2271-cba5-43e4-a179-7fd4389c6531\") " pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.579569 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-http" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.579615 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-grpc" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.579864 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-loki-s3" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.602368 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm"] Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.647636 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.707165 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt"] Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.723437 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.723517 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-config\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.723612 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.723733 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.723801 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llnjp\" (UniqueName: \"kubernetes.io/projected/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-kube-api-access-llnjp\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.723826 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.725414 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt"] Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.725526 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.750281 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-http" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.750995 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-grpc" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.826673 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfj75\" (UniqueName: \"kubernetes.io/projected/fbd5715c-12b1-4660-bdd6-39087b0ae32a-kube-api-access-vfj75\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.826836 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.826930 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.826995 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.827048 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llnjp\" (UniqueName: \"kubernetes.io/projected/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-kube-api-access-llnjp\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.827077 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.827113 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbd5715c-12b1-4660-bdd6-39087b0ae32a-config\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.827157 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.827199 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.827232 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.827279 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-config\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.828444 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.829497 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-config\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.833037 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.833778 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.850366 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llnjp\" (UniqueName: \"kubernetes.io/projected/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-kube-api-access-llnjp\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.855663 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/a9df4fe8-4ed5-4bd9-a514-9ed4259283b1-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-5467947bf7-dxbfm\" (UID: \"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1\") " pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.907879 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc"] Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.916274 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.921424 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway-ca-bundle" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.924101 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.924311 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.924469 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-http" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.924627 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.924828 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-client-http" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.929071 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.929188 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbd5715c-12b1-4660-bdd6-39087b0ae32a-config\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.929239 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.929300 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.929341 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfj75\" (UniqueName: \"kubernetes.io/projected/fbd5715c-12b1-4660-bdd6-39087b0ae32a-kube-api-access-vfj75\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.930671 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.931533 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbd5715c-12b1-4660-bdd6-39087b0ae32a-config\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.936737 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.936971 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-dockercfg-tc9sz" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.945199 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/fbd5715c-12b1-4660-bdd6-39087b0ae32a-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.949761 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.960161 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c"] Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.962099 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfj75\" (UniqueName: \"kubernetes.io/projected/fbd5715c-12b1-4660-bdd6-39087b0ae32a-kube-api-access-vfj75\") pod \"cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt\" (UID: \"fbd5715c-12b1-4660-bdd6-39087b0ae32a\") " pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.965669 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:12 crc kubenswrapper[4706]: I1208 19:40:12.988770 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc"] Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.005573 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c"] Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.038636 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.038735 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.038768 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tenants\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.038896 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.038974 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7gh8\" (UniqueName: \"kubernetes.io/projected/7811fdf0-d557-47cb-b700-4a072602f8c7-kube-api-access-w7gh8\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039039 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039089 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039120 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tls-secret\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039180 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tls-secret\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039249 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039331 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039367 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039438 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-rbac\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039533 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-rbac\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039765 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039809 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.039865 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tenants\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.040040 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ksj6\" (UniqueName: \"kubernetes.io/projected/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-kube-api-access-6ksj6\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.076982 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142538 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142604 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142639 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tenants\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142674 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ksj6\" (UniqueName: \"kubernetes.io/projected/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-kube-api-access-6ksj6\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142703 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142720 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142737 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tenants\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142765 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142787 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7gh8\" (UniqueName: \"kubernetes.io/projected/7811fdf0-d557-47cb-b700-4a072602f8c7-kube-api-access-w7gh8\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142814 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142840 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142865 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tls-secret\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142887 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tls-secret\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142910 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142936 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142965 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.142993 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-rbac\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.143023 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-rbac\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: E1208 19:40:13.143927 4706 secret.go:188] Couldn't get secret openstack/cloudkitty-lokistack-gateway-http: secret "cloudkitty-lokistack-gateway-http" not found Dec 08 19:40:13 crc kubenswrapper[4706]: E1208 19:40:13.144057 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tls-secret podName:7811fdf0-d557-47cb-b700-4a072602f8c7 nodeName:}" failed. No retries permitted until 2025-12-08 19:40:13.644028316 +0000 UTC m=+1096.286229319 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tls-secret") pod "cloudkitty-lokistack-gateway-bc75944f-m4nfc" (UID: "7811fdf0-d557-47cb-b700-4a072602f8c7") : secret "cloudkitty-lokistack-gateway-http" not found Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.144094 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.144133 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: E1208 19:40:13.144248 4706 secret.go:188] Couldn't get secret openstack/cloudkitty-lokistack-gateway-http: secret "cloudkitty-lokistack-gateway-http" not found Dec 08 19:40:13 crc kubenswrapper[4706]: E1208 19:40:13.144350 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tls-secret podName:d9d23ed1-9384-4bf8-adeb-2bc3340279ed nodeName:}" failed. No retries permitted until 2025-12-08 19:40:13.644304544 +0000 UTC m=+1096.286505547 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tls-secret") pod "cloudkitty-lokistack-gateway-bc75944f-7tk8c" (UID: "d9d23ed1-9384-4bf8-adeb-2bc3340279ed") : secret "cloudkitty-lokistack-gateway-http" not found Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.144572 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-rbac\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.144747 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.145439 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.145497 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.145504 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.145581 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.145643 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/7811fdf0-d557-47cb-b700-4a072602f8c7-rbac\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.146202 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.150907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.155311 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tenants\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.155872 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.156356 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tenants\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.169991 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ksj6\" (UniqueName: \"kubernetes.io/projected/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-kube-api-access-6ksj6\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.170819 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7gh8\" (UniqueName: \"kubernetes.io/projected/7811fdf0-d557-47cb-b700-4a072602f8c7-kube-api-access-w7gh8\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.531348 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.532660 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.536914 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-http" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.537077 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-grpc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.549938 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.654987 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655079 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tls-secret\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655107 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tls-secret\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655129 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655171 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv8k9\" (UniqueName: \"kubernetes.io/projected/9a912025-8f8e-4d61-861e-c933dafe4660-kube-api-access-wv8k9\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655197 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655234 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655277 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655313 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.655342 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a912025-8f8e-4d61-861e-c933dafe4660-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.660296 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/7811fdf0-d557-47cb-b700-4a072602f8c7-tls-secret\") pod \"cloudkitty-lokistack-gateway-bc75944f-m4nfc\" (UID: \"7811fdf0-d557-47cb-b700-4a072602f8c7\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.662549 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/d9d23ed1-9384-4bf8-adeb-2bc3340279ed-tls-secret\") pod \"cloudkitty-lokistack-gateway-bc75944f-7tk8c\" (UID: \"d9d23ed1-9384-4bf8-adeb-2bc3340279ed\") " pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.673564 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.676726 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.680236 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-grpc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.680643 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-http" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.682933 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.756875 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a912025-8f8e-4d61-861e-c933dafe4660-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.756919 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757000 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757033 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757066 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757089 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757116 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdnvq\" (UniqueName: \"kubernetes.io/projected/a7e3aaf0-5179-4837-b6ca-53276d637d10-kube-api-access-xdnvq\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757140 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv8k9\" (UniqueName: \"kubernetes.io/projected/9a912025-8f8e-4d61-861e-c933dafe4660-kube-api-access-wv8k9\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757164 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757197 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757219 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757240 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757273 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757291 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e3aaf0-5179-4837-b6ca-53276d637d10-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.757486 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.758159 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.758295 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a912025-8f8e-4d61-861e-c933dafe4660-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.758804 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.758183 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.761511 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.764588 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.773194 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/9a912025-8f8e-4d61-861e-c933dafe4660-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.783813 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv8k9\" (UniqueName: \"kubernetes.io/projected/9a912025-8f8e-4d61-861e-c933dafe4660-kube-api-access-wv8k9\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.801026 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.812039 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"9a912025-8f8e-4d61-861e-c933dafe4660\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.828532 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.830340 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.837807 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.849995 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-http" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.850234 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-grpc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.862062 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.874419 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdnvq\" (UniqueName: \"kubernetes.io/projected/a7e3aaf0-5179-4837-b6ca-53276d637d10-kube-api-access-xdnvq\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.874606 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.874655 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.874712 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e3aaf0-5179-4837-b6ca-53276d637d10-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.874816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.875043 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.875161 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.877124 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.890159 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.890310 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.890348 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e3aaf0-5179-4837-b6ca-53276d637d10-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.891440 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.891950 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.897977 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.903684 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/a7e3aaf0-5179-4837-b6ca-53276d637d10-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.915232 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.926653 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdnvq\" (UniqueName: \"kubernetes.io/projected/a7e3aaf0-5179-4837-b6ca-53276d637d10-kube-api-access-xdnvq\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"a7e3aaf0-5179-4837-b6ca-53276d637d10\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.976651 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.977011 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.977130 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1845b4-bbc3-4661-83d4-72b972fa288a-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.977308 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf5pc\" (UniqueName: \"kubernetes.io/projected/4d1845b4-bbc3-4661-83d4-72b972fa288a-kube-api-access-lf5pc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.977406 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.977498 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:13 crc kubenswrapper[4706]: I1208 19:40:13.977601 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.047498 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.078936 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf5pc\" (UniqueName: \"kubernetes.io/projected/4d1845b4-bbc3-4661-83d4-72b972fa288a-kube-api-access-lf5pc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.078988 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.079020 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.079057 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.079108 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.079130 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.079157 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1845b4-bbc3-4661-83d4-72b972fa288a-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.080029 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1845b4-bbc3-4661-83d4-72b972fa288a-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.085135 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.085273 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.091489 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.096990 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.108601 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/4d1845b4-bbc3-4661-83d4-72b972fa288a-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.118122 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf5pc\" (UniqueName: \"kubernetes.io/projected/4d1845b4-bbc3-4661-83d4-72b972fa288a-kube-api-access-lf5pc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.118950 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"4d1845b4-bbc3-4661-83d4-72b972fa288a\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:14 crc kubenswrapper[4706]: I1208 19:40:14.207712 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:20 crc kubenswrapper[4706]: I1208 19:40:20.025070 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.523555 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.524141 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5fwh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-q5fhn_openstack(dd4d903d-5735-40d1-9131-e587537401cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.525358 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" podUID="dd4d903d-5735-40d1-9131-e587537401cf" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.553966 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.554167 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nmgtv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-9kbjh_openstack(b9f26f6c-226a-4026-bde7-10f4401f53e0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.555738 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.563370 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.563560 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dp7jc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-n2vgp_openstack(acfc18e7-14e4-4aff-b304-b293233bb76e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.564720 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" podUID="acfc18e7-14e4-4aff-b304-b293233bb76e" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.587916 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.589674 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qz9j6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-lqbnv_openstack(52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.590892 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.833587 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" Dec 08 19:40:21 crc kubenswrapper[4706]: E1208 19:40:21.841018 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.259106 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.323522 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.328331 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5fwh\" (UniqueName: \"kubernetes.io/projected/dd4d903d-5735-40d1-9131-e587537401cf-kube-api-access-x5fwh\") pod \"dd4d903d-5735-40d1-9131-e587537401cf\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.328392 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-config\") pod \"dd4d903d-5735-40d1-9131-e587537401cf\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.328473 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-dns-svc\") pod \"dd4d903d-5735-40d1-9131-e587537401cf\" (UID: \"dd4d903d-5735-40d1-9131-e587537401cf\") " Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.329713 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-config" (OuterVolumeSpecName: "config") pod "dd4d903d-5735-40d1-9131-e587537401cf" (UID: "dd4d903d-5735-40d1-9131-e587537401cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.330184 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd4d903d-5735-40d1-9131-e587537401cf" (UID: "dd4d903d-5735-40d1-9131-e587537401cf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.334416 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd4d903d-5735-40d1-9131-e587537401cf-kube-api-access-x5fwh" (OuterVolumeSpecName: "kube-api-access-x5fwh") pod "dd4d903d-5735-40d1-9131-e587537401cf" (UID: "dd4d903d-5735-40d1-9131-e587537401cf"). InnerVolumeSpecName "kube-api-access-x5fwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.432497 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp7jc\" (UniqueName: \"kubernetes.io/projected/acfc18e7-14e4-4aff-b304-b293233bb76e-kube-api-access-dp7jc\") pod \"acfc18e7-14e4-4aff-b304-b293233bb76e\" (UID: \"acfc18e7-14e4-4aff-b304-b293233bb76e\") " Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.432575 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acfc18e7-14e4-4aff-b304-b293233bb76e-config\") pod \"acfc18e7-14e4-4aff-b304-b293233bb76e\" (UID: \"acfc18e7-14e4-4aff-b304-b293233bb76e\") " Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.433245 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5fwh\" (UniqueName: \"kubernetes.io/projected/dd4d903d-5735-40d1-9131-e587537401cf-kube-api-access-x5fwh\") on node \"crc\" DevicePath \"\"" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.433294 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.433312 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd4d903d-5735-40d1-9131-e587537401cf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.434361 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acfc18e7-14e4-4aff-b304-b293233bb76e-config" (OuterVolumeSpecName: "config") pod "acfc18e7-14e4-4aff-b304-b293233bb76e" (UID: "acfc18e7-14e4-4aff-b304-b293233bb76e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.440854 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acfc18e7-14e4-4aff-b304-b293233bb76e-kube-api-access-dp7jc" (OuterVolumeSpecName: "kube-api-access-dp7jc") pod "acfc18e7-14e4-4aff-b304-b293233bb76e" (UID: "acfc18e7-14e4-4aff-b304-b293233bb76e"). InnerVolumeSpecName "kube-api-access-dp7jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.535859 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acfc18e7-14e4-4aff-b304-b293233bb76e-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.535897 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp7jc\" (UniqueName: \"kubernetes.io/projected/acfc18e7-14e4-4aff-b304-b293233bb76e-kube-api-access-dp7jc\") on node \"crc\" DevicePath \"\"" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.776763 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.851658 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"13532f07-b125-4d67-bdb8-006b583bea40","Type":"ContainerStarted","Data":"5eb2f13c55231b3b6cb010d0eb1b7670de2631b0f253d0917078fd2c47f9c233"} Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.854040 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c","Type":"ContainerStarted","Data":"687255fab69e67a996ea975141fefb12cc36b67b225c7e72f904c247b9892256"} Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.856385 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.856391 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-q5fhn" event={"ID":"dd4d903d-5735-40d1-9131-e587537401cf","Type":"ContainerDied","Data":"4d92f7fd0d0660140c731a5108b834158909273407288d582cec2fb1730e857e"} Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.858129 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" event={"ID":"acfc18e7-14e4-4aff-b304-b293233bb76e","Type":"ContainerDied","Data":"9a58bc443407e1958b49f1aa31c2a7ac05dc6a8971cf29f91f003703f1672d70"} Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.858175 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-n2vgp" Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.953142 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q5fhn"] Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.966404 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q5fhn"] Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.983328 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n2vgp"] Dec 08 19:40:23 crc kubenswrapper[4706]: I1208 19:40:23.988851 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n2vgp"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.196137 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.202541 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.209075 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.215206 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081f2271_cba5_43e4_a179_7fd4389c6531.slice/crio-57e0efea9b2e97d1a96ae42b9c90a383ab8ab06f1514cf21c1fda107f86a795f WatchSource:0}: Error finding container 57e0efea9b2e97d1a96ae42b9c90a383ab8ab06f1514cf21c1fda107f86a795f: Status 404 returned error can't find the container with id 57e0efea9b2e97d1a96ae42b9c90a383ab8ab06f1514cf21c1fda107f86a795f Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.228746 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd927fec_2b54_41d2_affb_19f9fd8c96ab.slice/crio-2c196f4cf26981fb957b49a15f6d2b48aef3a031ff9e769d4efb9d1d6a6513ba WatchSource:0}: Error finding container 2c196f4cf26981fb957b49a15f6d2b48aef3a031ff9e769d4efb9d1d6a6513ba: Status 404 returned error can't find the container with id 2c196f4cf26981fb957b49a15f6d2b48aef3a031ff9e769d4efb9d1d6a6513ba Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.408804 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.412853 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod908bed13_e8ef_4234_83fd_21b3f0ed218e.slice/crio-d50e959f83cc5259cb909c1a186034e36ca0c871fe97b949cd978ebd975e868d WatchSource:0}: Error finding container d50e959f83cc5259cb909c1a186034e36ca0c871fe97b949cd978ebd975e868d: Status 404 returned error can't find the container with id d50e959f83cc5259cb909c1a186034e36ca0c871fe97b949cd978ebd975e868d Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.419569 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.440784 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78wps"] Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.508218 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4853c157_69b5_45ec_8fad_9e5ddea58cea.slice/crio-8e4155b204c3112d62364734501d944668cbef0cd75413512668b584761b6e12 WatchSource:0}: Error finding container 8e4155b204c3112d62364734501d944668cbef0cd75413512668b584761b6e12: Status 404 returned error can't find the container with id 8e4155b204c3112d62364734501d944668cbef0cd75413512668b584761b6e12 Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.512122 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79f551cd_d08d_4739_9041_4543f05fbb60.slice/crio-41e83e3f81b2a432b17234af29f1f9cd5508c8f36d894ff70ed9dbf5a8e0e946 WatchSource:0}: Error finding container 41e83e3f81b2a432b17234af29f1f9cd5508c8f36d894ff70ed9dbf5a8e0e946: Status 404 returned error can't find the container with id 41e83e3f81b2a432b17234af29f1f9cd5508c8f36d894ff70ed9dbf5a8e0e946 Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.559716 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.597035 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.616891 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.629930 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.677816 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.809031 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7811fdf0_d557_47cb_b700_4a072602f8c7.slice/crio-1f2d8f8fa7fec702f03b2771bd54d8bcf3abd0c13300ea469ac52946479857cc WatchSource:0}: Error finding container 1f2d8f8fa7fec702f03b2771bd54d8bcf3abd0c13300ea469ac52946479857cc: Status 404 returned error can't find the container with id 1f2d8f8fa7fec702f03b2771bd54d8bcf3abd0c13300ea469ac52946479857cc Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.849995 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.860702 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.868490 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.871970 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fd927fec-2b54-41d2-affb-19f9fd8c96ab","Type":"ContainerStarted","Data":"2c196f4cf26981fb957b49a15f6d2b48aef3a031ff9e769d4efb9d1d6a6513ba"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.873373 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f498bda7-c379-4808-bd6c-8e5513c4322d","Type":"ContainerStarted","Data":"4135b0148c9c5c0fe4eafe79dcc3a07d217a2a236191717a6e0572fdee084890"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.875341 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" event={"ID":"7811fdf0-d557-47cb-b700-4a072602f8c7","Type":"ContainerStarted","Data":"1f2d8f8fa7fec702f03b2771bd54d8bcf3abd0c13300ea469ac52946479857cc"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.877212 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c"] Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.877313 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"9a912025-8f8e-4d61-861e-c933dafe4660","Type":"ContainerStarted","Data":"1a99899768aa7fc989bac5ee805ad5e95f8fdfe868017819ccb1350f135810ec"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.879019 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerStarted","Data":"8e4155b204c3112d62364734501d944668cbef0cd75413512668b584761b6e12"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.880288 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" event={"ID":"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1","Type":"ContainerStarted","Data":"47bd424a8cf443a1ab65f30d3a11303a28207a8cdec8146fbf1bfcabde3efb8f"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.881450 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" event={"ID":"081f2271-cba5-43e4-a179-7fd4389c6531","Type":"ContainerStarted","Data":"57e0efea9b2e97d1a96ae42b9c90a383ab8ab06f1514cf21c1fda107f86a795f"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.882816 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88233d80-672d-47c3-9a26-70c22f748e29","Type":"ContainerStarted","Data":"8d5bdf335a2dc4b208339495f3b87b35f90199835f6469d1f652ccada07bd114"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.884342 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78wps" event={"ID":"79f551cd-d08d-4739-9041-4543f05fbb60","Type":"ContainerStarted","Data":"41e83e3f81b2a432b17234af29f1f9cd5508c8f36d894ff70ed9dbf5a8e0e946"} Dec 08 19:40:24 crc kubenswrapper[4706]: I1208 19:40:24.885504 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"908bed13-e8ef-4234-83fd-21b3f0ed218e","Type":"ContainerStarted","Data":"d50e959f83cc5259cb909c1a186034e36ca0c871fe97b949cd978ebd975e868d"} Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.911234 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40f5d973_ac12_469a_a1d9_aa871b7cc9a3.slice/crio-a2dd17efbc25a58d0dd5254a7d668ee97d4d247e2be0777d83ca324c10aa9358 WatchSource:0}: Error finding container a2dd17efbc25a58d0dd5254a7d668ee97d4d247e2be0777d83ca324c10aa9358: Status 404 returned error can't find the container with id a2dd17efbc25a58d0dd5254a7d668ee97d4d247e2be0777d83ca324c10aa9358 Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.915241 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d1845b4_bbc3_4661_83d4_72b972fa288a.slice/crio-a0b078e342700fc75a021e82c5224b9a79ce4225cce710f5bcb2b20cc89c70f3 WatchSource:0}: Error finding container a0b078e342700fc75a021e82c5224b9a79ce4225cce710f5bcb2b20cc89c70f3: Status 404 returned error can't find the container with id a0b078e342700fc75a021e82c5224b9a79ce4225cce710f5bcb2b20cc89c70f3 Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.918946 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7e3aaf0_5179_4837_b6ca_53276d637d10.slice/crio-20d12f818f2a5c320c6323a1c3189f24a377ef5decb558c8ee3a846443ae3c7a WatchSource:0}: Error finding container 20d12f818f2a5c320c6323a1c3189f24a377ef5decb558c8ee3a846443ae3c7a: Status 404 returned error can't find the container with id 20d12f818f2a5c320c6323a1c3189f24a377ef5decb558c8ee3a846443ae3c7a Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.922632 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbd5715c_12b1_4660_bdd6_39087b0ae32a.slice/crio-24061dbe8b29c816d024d37b8d6646dc50a722fd93d60adde03eae515be98c8b WatchSource:0}: Error finding container 24061dbe8b29c816d024d37b8d6646dc50a722fd93d60adde03eae515be98c8b: Status 404 returned error can't find the container with id 24061dbe8b29c816d024d37b8d6646dc50a722fd93d60adde03eae515be98c8b Dec 08 19:40:24 crc kubenswrapper[4706]: W1208 19:40:24.927016 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9d23ed1_9384_4bf8_adeb_2bc3340279ed.slice/crio-a924070fb5b8caffdbf4ea1de361bca077969fdf63204d5e852414b6e4fff91d WatchSource:0}: Error finding container a924070fb5b8caffdbf4ea1de361bca077969fdf63204d5e852414b6e4fff91d: Status 404 returned error can't find the container with id a924070fb5b8caffdbf4ea1de361bca077969fdf63204d5e852414b6e4fff91d Dec 08 19:40:24 crc kubenswrapper[4706]: E1208 19:40:24.930643 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:gateway,Image:registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:129dbfaf84e687adb93f670d2b46754fd2562513f6a45f79b37c7cc4c622f53e,Command:[],Args:[--debug.name=lokistack-gateway --web.listen=0.0.0.0:8080 --web.internal.listen=0.0.0.0:8081 --web.healthchecks.url=https://localhost:8080 --log.level=warn --logs.read.endpoint=https://cloudkitty-lokistack-query-frontend-http.openstack.svc.cluster.local:3100 --logs.tail.endpoint=https://cloudkitty-lokistack-query-frontend-http.openstack.svc.cluster.local:3100 --logs.write.endpoint=https://cloudkitty-lokistack-distributor-http.openstack.svc.cluster.local:3100 --logs.write-timeout=4m0s --rbac.config=/etc/lokistack-gateway/rbac.yaml --tenants.config=/etc/lokistack-gateway/tenants.yaml --server.read-timeout=48s --server.write-timeout=6m0s --tls.min-version=VersionTLS12 --tls.server.cert-file=/var/run/tls/http/server/tls.crt --tls.server.key-file=/var/run/tls/http/server/tls.key --tls.healthchecks.server-ca-file=/var/run/ca/server/service-ca.crt --tls.healthchecks.server-name=cloudkitty-lokistack-gateway-http.openstack.svc.cluster.local --tls.internal.server.cert-file=/var/run/tls/http/server/tls.crt --tls.internal.server.key-file=/var/run/tls/http/server/tls.key --tls.min-version=VersionTLS12 --tls.cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --logs.tls.ca-file=/var/run/ca/upstream/service-ca.crt --logs.tls.cert-file=/var/run/tls/http/upstream/tls.crt --logs.tls.key-file=/var/run/tls/http/upstream/tls.key --tls.client-auth-type=RequestClientCert],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},ContainerPort{Name:public,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rbac,ReadOnly:true,MountPath:/etc/lokistack-gateway/rbac.yaml,SubPath:rbac.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tenants,ReadOnly:true,MountPath:/etc/lokistack-gateway/tenants.yaml,SubPath:tenants.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:lokistack-gateway,ReadOnly:true,MountPath:/etc/lokistack-gateway/lokistack-gateway.rego,SubPath:lokistack-gateway.rego,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tls-secret,ReadOnly:true,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-gateway-client-http,ReadOnly:true,MountPath:/var/run/tls/http/upstream,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:true,MountPath:/var/run/ca/upstream,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-gateway-ca-bundle,ReadOnly:true,MountPath:/var/run/ca/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-ca-bundle,ReadOnly:false,MountPath:/var/run/tenants-ca/cloudkitty,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ksj6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/live,Port:{0 8081 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 8081 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:12,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-gateway-bc75944f-7tk8c_openstack(d9d23ed1-9384-4bf8-adeb-2bc3340279ed): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:40:24 crc kubenswrapper[4706]: E1208 19:40:24.932609 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" podUID="d9d23ed1-9384-4bf8-adeb-2bc3340279ed" Dec 08 19:40:24 crc kubenswrapper[4706]: E1208 19:40:24.935990 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-query-frontend,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7,Command:[],Args:[-target=query-frontend -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-query-frontend-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-query-frontend-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vfj75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt_openstack(fbd5715c-12b1-4660-bdd6-39087b0ae32a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 19:40:24 crc kubenswrapper[4706]: E1208 19:40:24.937323 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-query-frontend\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" podUID="fbd5715c-12b1-4660-bdd6-39087b0ae32a" Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.530843 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fb6gx"] Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.638974 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acfc18e7-14e4-4aff-b304-b293233bb76e" path="/var/lib/kubelet/pods/acfc18e7-14e4-4aff-b304-b293233bb76e/volumes" Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.639611 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd4d903d-5735-40d1-9131-e587537401cf" path="/var/lib/kubelet/pods/dd4d903d-5735-40d1-9131-e587537401cf/volumes" Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.905716 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"4d1845b4-bbc3-4661-83d4-72b972fa288a","Type":"ContainerStarted","Data":"a0b078e342700fc75a021e82c5224b9a79ce4225cce710f5bcb2b20cc89c70f3"} Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.909983 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" event={"ID":"fbd5715c-12b1-4660-bdd6-39087b0ae32a","Type":"ContainerStarted","Data":"24061dbe8b29c816d024d37b8d6646dc50a722fd93d60adde03eae515be98c8b"} Dec 08 19:40:25 crc kubenswrapper[4706]: E1208 19:40:25.913045 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-query-frontend\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7\\\"\"" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" podUID="fbd5715c-12b1-4660-bdd6-39087b0ae32a" Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.920125 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ea8337d-cff3-414c-98f6-17b41653b4fb","Type":"ContainerStarted","Data":"4d2558eb0a6a323f3921e28819b25ee1b7d270e9ab07768a3bfcedfd8570cdbd"} Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.924483 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb6gx" event={"ID":"25287a9f-27c5-47a7-945e-684eea0157c8","Type":"ContainerStarted","Data":"807c32ddcdbf0c0e456ece5e8317c7df4741db460ee24f7343c244336ce77f9b"} Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.933019 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" event={"ID":"d9d23ed1-9384-4bf8-adeb-2bc3340279ed","Type":"ContainerStarted","Data":"a924070fb5b8caffdbf4ea1de361bca077969fdf63204d5e852414b6e4fff91d"} Dec 08 19:40:25 crc kubenswrapper[4706]: E1208 19:40:25.934826 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:129dbfaf84e687adb93f670d2b46754fd2562513f6a45f79b37c7cc4c622f53e\\\"\"" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" podUID="d9d23ed1-9384-4bf8-adeb-2bc3340279ed" Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.937646 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"a7e3aaf0-5179-4837-b6ca-53276d637d10","Type":"ContainerStarted","Data":"20d12f818f2a5c320c6323a1c3189f24a377ef5decb558c8ee3a846443ae3c7a"} Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.944396 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"40f5d973-ac12-469a-a1d9-aa871b7cc9a3","Type":"ContainerStarted","Data":"a2dd17efbc25a58d0dd5254a7d668ee97d4d247e2be0777d83ca324c10aa9358"} Dec 08 19:40:25 crc kubenswrapper[4706]: I1208 19:40:25.946790 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1472b8cb-795a-4a57-95f8-184a7707e8ad","Type":"ContainerStarted","Data":"b9240b7d3b9c39a64fdc169ca933af330b5f84a9bba65f7ebdcde2ee2fe3e519"} Dec 08 19:40:26 crc kubenswrapper[4706]: E1208 19:40:26.958419 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-query-frontend\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7\\\"\"" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" podUID="fbd5715c-12b1-4660-bdd6-39087b0ae32a" Dec 08 19:40:26 crc kubenswrapper[4706]: E1208 19:40:26.959003 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:129dbfaf84e687adb93f670d2b46754fd2562513f6a45f79b37c7cc4c622f53e\\\"\"" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" podUID="d9d23ed1-9384-4bf8-adeb-2bc3340279ed" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.686890 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.687914 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vt4p2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(156710b7-7dbe-4bb7-9c69-ce726d4b7a4c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.689163 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="156710b7-7dbe-4bb7-9c69-ce726d4b7a4c" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.694249 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.694670 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-index-gateway,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7,Command:[],Args:[-target=index-gateway -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:storage,ReadOnly:false,MountPath:/tmp/loki,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-index-gateway-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-index-gateway-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lf5pc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-index-gateway-0_openstack(4d1845b4-bbc3-4661-83d4-72b972fa288a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.696634 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-index-gateway\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" podUID="4d1845b4-bbc3-4661-83d4-72b972fa288a" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.698639 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.698735 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2nqxv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(13532f07-b125-4d67-bdb8-006b583bea40): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.699896 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="13532f07-b125-4d67-bdb8-006b583bea40" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.708638 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.708856 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-compactor,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7,Command:[],Args:[-target=compactor -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:storage,ReadOnly:false,MountPath:/tmp/loki,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-compactor-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-compactor-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xdnvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-compactor-0_openstack(a7e3aaf0-5179-4837-b6ca-53276d637d10): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.710160 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-compactor\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-compactor-0" podUID="a7e3aaf0-5179-4837-b6ca-53276d637d10" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.728243 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.728677 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-ingester,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7,Command:[],Args:[-target=ingester -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:gossip-ring,HostPort:0,ContainerPort:7946,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:storage,ReadOnly:false,MountPath:/tmp/loki,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:wal,ReadOnly:false,MountPath:/tmp/wal,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ingester-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ingester-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv8k9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-ingester-0_openstack(9a912025-8f8e-4d61-861e-c933dafe4660): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:40:38 crc kubenswrapper[4706]: E1208 19:40:38.729958 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-ingester\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="9a912025-8f8e-4d61-861e-c933dafe4660" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.086029 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-ingester\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7\\\"\"" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="9a912025-8f8e-4d61-861e-c933dafe4660" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.086029 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="13532f07-b125-4d67-bdb8-006b583bea40" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.086517 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-index-gateway\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7\\\"\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" podUID="4d1845b4-bbc3-4661-83d4-72b972fa288a" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.087147 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="156710b7-7dbe-4bb7-9c69-ce726d4b7a4c" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.090118 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-compactor\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:06b83c3cbf0c5db4dd9812e046ca14189d18cf7b3c7f2f2c37aa705cc5f5deb7\\\"\"" pod="openstack/cloudkitty-lokistack-compactor-0" podUID="a7e3aaf0-5179-4837-b6ca-53276d637d10" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.258716 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:129dbfaf84e687adb93f670d2b46754fd2562513f6a45f79b37c7cc4c622f53e" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.259068 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:gateway,Image:registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:129dbfaf84e687adb93f670d2b46754fd2562513f6a45f79b37c7cc4c622f53e,Command:[],Args:[--debug.name=lokistack-gateway --web.listen=0.0.0.0:8080 --web.internal.listen=0.0.0.0:8081 --web.healthchecks.url=https://localhost:8080 --log.level=warn --logs.read.endpoint=https://cloudkitty-lokistack-query-frontend-http.openstack.svc.cluster.local:3100 --logs.tail.endpoint=https://cloudkitty-lokistack-query-frontend-http.openstack.svc.cluster.local:3100 --logs.write.endpoint=https://cloudkitty-lokistack-distributor-http.openstack.svc.cluster.local:3100 --logs.write-timeout=4m0s --rbac.config=/etc/lokistack-gateway/rbac.yaml --tenants.config=/etc/lokistack-gateway/tenants.yaml --server.read-timeout=48s --server.write-timeout=6m0s --tls.min-version=VersionTLS12 --tls.server.cert-file=/var/run/tls/http/server/tls.crt --tls.server.key-file=/var/run/tls/http/server/tls.key --tls.healthchecks.server-ca-file=/var/run/ca/server/service-ca.crt --tls.healthchecks.server-name=cloudkitty-lokistack-gateway-http.openstack.svc.cluster.local --tls.internal.server.cert-file=/var/run/tls/http/server/tls.crt --tls.internal.server.key-file=/var/run/tls/http/server/tls.key --tls.min-version=VersionTLS12 --tls.cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --logs.tls.ca-file=/var/run/ca/upstream/service-ca.crt --logs.tls.cert-file=/var/run/tls/http/upstream/tls.crt --logs.tls.key-file=/var/run/tls/http/upstream/tls.key --tls.client-auth-type=RequestClientCert],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},ContainerPort{Name:public,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rbac,ReadOnly:true,MountPath:/etc/lokistack-gateway/rbac.yaml,SubPath:rbac.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tenants,ReadOnly:true,MountPath:/etc/lokistack-gateway/tenants.yaml,SubPath:tenants.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:lokistack-gateway,ReadOnly:true,MountPath:/etc/lokistack-gateway/lokistack-gateway.rego,SubPath:lokistack-gateway.rego,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tls-secret,ReadOnly:true,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-gateway-client-http,ReadOnly:true,MountPath:/var/run/tls/http/upstream,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:true,MountPath:/var/run/ca/upstream,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-gateway-ca-bundle,ReadOnly:true,MountPath:/var/run/ca/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-ca-bundle,ReadOnly:false,MountPath:/var/run/tenants-ca/cloudkitty,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w7gh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/live,Port:{0 8081 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 8081 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:12,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-gateway-bc75944f-m4nfc_openstack(7811fdf0-d557-47cb-b700-4a072602f8c7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.260316 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" podUID="7811fdf0-d557-47cb-b700-4a072602f8c7" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.514139 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.514424 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n87hb9h5fh597h5bfhffh595h644h4h557h54hc7h5dfh56bh8bh5bfh5b6hcfh5cchb9h5d8h54ch65bh4h675h89h5fdh6ch645h58fh56h5fbq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5cvx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(40f5d973-ac12-469a-a1d9-aa871b7cc9a3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.828228 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.828978 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54fh695h5b6hdh5c7h65fhffh97h5b4h7bh55h5ddh94h64ch88h59fhc7hf7h569h58dh74h679h697h9bh5fbhdbh58dhdch67dh594hf5h5d4q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l7c2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-78wps_openstack(79f551cd-d08d-4739-9041-4543f05fbb60): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:39 crc kubenswrapper[4706]: E1208 19:40:39.830728 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-78wps" podUID="79f551cd-d08d-4739-9041-4543f05fbb60" Dec 08 19:40:40 crc kubenswrapper[4706]: E1208 19:40:40.037984 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Dec 08 19:40:40 crc kubenswrapper[4706]: E1208 19:40:40.038240 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n644h567hc9h96h67dh599h554h684h58fh575h596hfdh5c7h66hc8hf9h5d6h5c9h54ch5h644hfch5cfh8bh558h64fh67fh598h67fh588h5cdh68cq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dhbfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(f498bda7-c379-4808-bd6c-8e5513c4322d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:40:40 crc kubenswrapper[4706]: E1208 19:40:40.094005 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:129dbfaf84e687adb93f670d2b46754fd2562513f6a45f79b37c7cc4c622f53e\\\"\"" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" podUID="7811fdf0-d557-47cb-b700-4a072602f8c7" Dec 08 19:40:40 crc kubenswrapper[4706]: E1208 19:40:40.094005 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-78wps" podUID="79f551cd-d08d-4739-9041-4543f05fbb60" Dec 08 19:40:41 crc kubenswrapper[4706]: E1208 19:40:41.960496 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 08 19:40:41 crc kubenswrapper[4706]: E1208 19:40:41.961093 4706 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 08 19:40:41 crc kubenswrapper[4706]: E1208 19:40:41.961350 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8s94p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(908bed13-e8ef-4234-83fd-21b3f0ed218e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 19:40:41 crc kubenswrapper[4706]: E1208 19:40:41.962737 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="908bed13-e8ef-4234-83fd-21b3f0ed218e" Dec 08 19:40:42 crc kubenswrapper[4706]: E1208 19:40:42.112904 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="908bed13-e8ef-4234-83fd-21b3f0ed218e" Dec 08 19:40:45 crc kubenswrapper[4706]: I1208 19:40:45.145795 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fd927fec-2b54-41d2-affb-19f9fd8c96ab","Type":"ContainerStarted","Data":"a15dba23a5985de22035f864b1db378fb763b1f40b059fa3b9c71d226962963c"} Dec 08 19:40:45 crc kubenswrapper[4706]: I1208 19:40:45.146613 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 08 19:40:45 crc kubenswrapper[4706]: I1208 19:40:45.171307 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=31.936562975 podStartE2EDuration="47.171285653s" podCreationTimestamp="2025-12-08 19:39:58 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.238776085 +0000 UTC m=+1106.880977088" lastFinishedPulling="2025-12-08 19:40:39.473498763 +0000 UTC m=+1122.115699766" observedRunningTime="2025-12-08 19:40:45.168940706 +0000 UTC m=+1127.811141709" watchObservedRunningTime="2025-12-08 19:40:45.171285653 +0000 UTC m=+1127.813486656" Dec 08 19:40:45 crc kubenswrapper[4706]: E1208 19:40:45.569196 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="f498bda7-c379-4808-bd6c-8e5513c4322d" Dec 08 19:40:45 crc kubenswrapper[4706]: E1208 19:40:45.572033 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="40f5d973-ac12-469a-a1d9-aa871b7cc9a3" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.158328 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" event={"ID":"fbd5715c-12b1-4660-bdd6-39087b0ae32a","Type":"ContainerStarted","Data":"7324e9a8d3e30a4625f382eaa2fcfb93ae9239278a6733d2be30e1655add13a0"} Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.158715 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.160899 4706 generic.go:334] "Generic (PLEG): container finished" podID="b9f26f6c-226a-4026-bde7-10f4401f53e0" containerID="0e36fae90b50af38a0c7268fc57d4f4210ccca86adffa06d94f3f473b19c6114" exitCode=0 Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.160963 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" event={"ID":"b9f26f6c-226a-4026-bde7-10f4401f53e0","Type":"ContainerDied","Data":"0e36fae90b50af38a0c7268fc57d4f4210ccca86adffa06d94f3f473b19c6114"} Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.163491 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" event={"ID":"081f2271-cba5-43e4-a179-7fd4389c6531","Type":"ContainerStarted","Data":"6956b98d321c99d8bd9eb27b2fcc16da6cf2edd2f46bc62274c2fcfdd251f210"} Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.163689 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.165555 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" event={"ID":"a9df4fe8-4ed5-4bd9-a514-9ed4259283b1","Type":"ContainerStarted","Data":"87a042501ff70d73aab1c0094519130de20083d8eb86307f8ffba136c8c397d8"} Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.165738 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.167283 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" event={"ID":"d9d23ed1-9384-4bf8-adeb-2bc3340279ed","Type":"ContainerStarted","Data":"05417e7e20c75b6e715644364d53c92032ff57d082de4c5966b9ab42137a073a"} Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.167713 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.169321 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f498bda7-c379-4808-bd6c-8e5513c4322d","Type":"ContainerStarted","Data":"7708a1ed054146dfd5ff04ceea0a11ee6b77f75c2525d4b92c222908bfcca503"} Dec 08 19:40:46 crc kubenswrapper[4706]: E1208 19:40:46.171048 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="f498bda7-c379-4808-bd6c-8e5513c4322d" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.171512 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb6gx" event={"ID":"25287a9f-27c5-47a7-945e-684eea0157c8","Type":"ContainerStarted","Data":"da66b6465280ebb35976997547e030733cab455a0418232223379b133fc022ba"} Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.174316 4706 generic.go:334] "Generic (PLEG): container finished" podID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" containerID="f4312efa40125432255026a9fc1254f221ccfdb2a54152a5d299b4351291faa2" exitCode=0 Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.174386 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" event={"ID":"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568","Type":"ContainerDied","Data":"f4312efa40125432255026a9fc1254f221ccfdb2a54152a5d299b4351291faa2"} Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.185319 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"40f5d973-ac12-469a-a1d9-aa871b7cc9a3","Type":"ContainerStarted","Data":"0fde3e62a370aff89328d278477a14b78d1725484ab2cd1fc113e9b46e3a71b2"} Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.186884 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" podStartSLOduration=17.55088299 podStartE2EDuration="34.186859305s" podCreationTimestamp="2025-12-08 19:40:12 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.935821562 +0000 UTC m=+1107.578022565" lastFinishedPulling="2025-12-08 19:40:41.571797877 +0000 UTC m=+1124.213998880" observedRunningTime="2025-12-08 19:40:46.179863566 +0000 UTC m=+1128.822064579" watchObservedRunningTime="2025-12-08 19:40:46.186859305 +0000 UTC m=+1128.829060308" Dec 08 19:40:46 crc kubenswrapper[4706]: E1208 19:40:46.188029 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="40f5d973-ac12-469a-a1d9-aa871b7cc9a3" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.195320 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.214535 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" podStartSLOduration=19.033881693 podStartE2EDuration="34.214504031s" podCreationTimestamp="2025-12-08 19:40:12 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.633389145 +0000 UTC m=+1107.275590138" lastFinishedPulling="2025-12-08 19:40:39.814011473 +0000 UTC m=+1122.456212476" observedRunningTime="2025-12-08 19:40:46.203627482 +0000 UTC m=+1128.845828485" watchObservedRunningTime="2025-12-08 19:40:46.214504031 +0000 UTC m=+1128.856705054" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.298647 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" podStartSLOduration=18.062922502 podStartE2EDuration="34.298615495s" podCreationTimestamp="2025-12-08 19:40:12 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.231412856 +0000 UTC m=+1106.873613859" lastFinishedPulling="2025-12-08 19:40:40.467105849 +0000 UTC m=+1123.109306852" observedRunningTime="2025-12-08 19:40:46.286964114 +0000 UTC m=+1128.929165117" watchObservedRunningTime="2025-12-08 19:40:46.298615495 +0000 UTC m=+1128.940816498" Dec 08 19:40:46 crc kubenswrapper[4706]: I1208 19:40:46.339276 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-7tk8c" podStartSLOduration=15.855985826 podStartE2EDuration="34.33920891s" podCreationTimestamp="2025-12-08 19:40:12 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.930318236 +0000 UTC m=+1107.572519239" lastFinishedPulling="2025-12-08 19:40:43.41354131 +0000 UTC m=+1126.055742323" observedRunningTime="2025-12-08 19:40:46.337492331 +0000 UTC m=+1128.979693354" watchObservedRunningTime="2025-12-08 19:40:46.33920891 +0000 UTC m=+1128.981409923" Dec 08 19:40:47 crc kubenswrapper[4706]: I1208 19:40:47.197828 4706 generic.go:334] "Generic (PLEG): container finished" podID="25287a9f-27c5-47a7-945e-684eea0157c8" containerID="da66b6465280ebb35976997547e030733cab455a0418232223379b133fc022ba" exitCode=0 Dec 08 19:40:47 crc kubenswrapper[4706]: I1208 19:40:47.197956 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb6gx" event={"ID":"25287a9f-27c5-47a7-945e-684eea0157c8","Type":"ContainerDied","Data":"da66b6465280ebb35976997547e030733cab455a0418232223379b133fc022ba"} Dec 08 19:40:47 crc kubenswrapper[4706]: I1208 19:40:47.200785 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerStarted","Data":"e2c9357f543311893043863bba13f3ac1a2ce573ff909d0205d3d74014d39082"} Dec 08 19:40:47 crc kubenswrapper[4706]: E1208 19:40:47.203517 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="f498bda7-c379-4808-bd6c-8e5513c4322d" Dec 08 19:40:47 crc kubenswrapper[4706]: E1208 19:40:47.204021 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="40f5d973-ac12-469a-a1d9-aa871b7cc9a3" Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.212153 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" event={"ID":"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568","Type":"ContainerStarted","Data":"ae486bf81a5b12911f8587408d3868244a134566a5e315c09dac14d9c7bff8e9"} Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.212796 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.217161 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb6gx" event={"ID":"25287a9f-27c5-47a7-945e-684eea0157c8","Type":"ContainerStarted","Data":"3c9d7f32b3fa4c1c12f96a1c6a8a29d424a1f5e930a220a321a0599868b0313f"} Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.217615 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.217637 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb6gx" event={"ID":"25287a9f-27c5-47a7-945e-684eea0157c8","Type":"ContainerStarted","Data":"9cee1f19fb60d47d17d69a1864348035773666178d263ad3d3da123cc3bca742"} Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.217654 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.219953 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" event={"ID":"b9f26f6c-226a-4026-bde7-10f4401f53e0","Type":"ContainerStarted","Data":"3f95beea4e2b84fc7411a2303591cb075e6e10023ff74e91a2f2fd080473c7e0"} Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.220230 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.222091 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88233d80-672d-47c3-9a26-70c22f748e29","Type":"ContainerStarted","Data":"3bec0e62246a8700e5ec85066fab5b59dc7b4d9c6faed38cabb0e8a5059162bc"} Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.246091 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" podStartSLOduration=7.749099842 podStartE2EDuration="54.246062726s" podCreationTimestamp="2025-12-08 19:39:54 +0000 UTC" firstStartedPulling="2025-12-08 19:39:55.074837913 +0000 UTC m=+1077.717038916" lastFinishedPulling="2025-12-08 19:40:41.571800797 +0000 UTC m=+1124.214001800" observedRunningTime="2025-12-08 19:40:48.235490315 +0000 UTC m=+1130.877691318" watchObservedRunningTime="2025-12-08 19:40:48.246062726 +0000 UTC m=+1130.888263729" Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.299139 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-fb6gx" podStartSLOduration=29.833130626 podStartE2EDuration="44.299110086s" podCreationTimestamp="2025-12-08 19:40:04 +0000 UTC" firstStartedPulling="2025-12-08 19:40:25.672446895 +0000 UTC m=+1108.314647888" lastFinishedPulling="2025-12-08 19:40:40.138426345 +0000 UTC m=+1122.780627348" observedRunningTime="2025-12-08 19:40:48.291234871 +0000 UTC m=+1130.933435894" watchObservedRunningTime="2025-12-08 19:40:48.299110086 +0000 UTC m=+1130.941311089" Dec 08 19:40:48 crc kubenswrapper[4706]: I1208 19:40:48.315380 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" podStartSLOduration=8.505810518 podStartE2EDuration="55.315357178s" podCreationTimestamp="2025-12-08 19:39:53 +0000 UTC" firstStartedPulling="2025-12-08 19:39:54.76235727 +0000 UTC m=+1077.404558273" lastFinishedPulling="2025-12-08 19:40:41.57190392 +0000 UTC m=+1124.214104933" observedRunningTime="2025-12-08 19:40:48.311427826 +0000 UTC m=+1130.953628859" watchObservedRunningTime="2025-12-08 19:40:48.315357178 +0000 UTC m=+1130.957558181" Dec 08 19:40:50 crc kubenswrapper[4706]: I1208 19:40:50.242501 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"9a912025-8f8e-4d61-861e-c933dafe4660","Type":"ContainerStarted","Data":"a6add835520d81e5621c0b4ce37d0a4cc1baaf0619fe5a3e00ca96cf3312e4c1"} Dec 08 19:40:50 crc kubenswrapper[4706]: I1208 19:40:50.243145 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:40:50 crc kubenswrapper[4706]: I1208 19:40:50.266838 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-ingester-0" podStartSLOduration=-9223371998.587965 podStartE2EDuration="38.266811234s" podCreationTimestamp="2025-12-08 19:40:12 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.600022606 +0000 UTC m=+1107.242223609" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:40:50.263072627 +0000 UTC m=+1132.905273630" watchObservedRunningTime="2025-12-08 19:40:50.266811234 +0000 UTC m=+1132.909012237" Dec 08 19:40:52 crc kubenswrapper[4706]: I1208 19:40:52.265979 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"13532f07-b125-4d67-bdb8-006b583bea40","Type":"ContainerStarted","Data":"c5e3902cfeb117ee9d279959c9f0d4dcd09139afb1f3bc7dac7186d6280903f5"} Dec 08 19:40:53 crc kubenswrapper[4706]: I1208 19:40:53.278804 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"a7e3aaf0-5179-4837-b6ca-53276d637d10","Type":"ContainerStarted","Data":"fef36ec8e32eb0ace47860f5e85915e9a9a3fe2302d879f8e1d90c4975bb4e9d"} Dec 08 19:40:53 crc kubenswrapper[4706]: I1208 19:40:53.279722 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:40:53 crc kubenswrapper[4706]: I1208 19:40:53.281768 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c","Type":"ContainerStarted","Data":"de2e8754471cd05408f23608f1b340aec92752980100888fff1b7851ac3d2124"} Dec 08 19:40:53 crc kubenswrapper[4706]: I1208 19:40:53.308370 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-compactor-0" podStartSLOduration=-9223371995.54644 podStartE2EDuration="41.30833629s" podCreationTimestamp="2025-12-08 19:40:12 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.926819326 +0000 UTC m=+1107.569020329" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:40:53.299126648 +0000 UTC m=+1135.941327651" watchObservedRunningTime="2025-12-08 19:40:53.30833629 +0000 UTC m=+1135.950537293" Dec 08 19:40:53 crc kubenswrapper[4706]: I1208 19:40:53.495809 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 08 19:40:53 crc kubenswrapper[4706]: I1208 19:40:53.975615 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.293297 4706 generic.go:334] "Generic (PLEG): container finished" podID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerID="e2c9357f543311893043863bba13f3ac1a2ce573ff909d0205d3d74014d39082" exitCode=0 Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.293378 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerDied","Data":"e2c9357f543311893043863bba13f3ac1a2ce573ff909d0205d3d74014d39082"} Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.296208 4706 generic.go:334] "Generic (PLEG): container finished" podID="88233d80-672d-47c3-9a26-70c22f748e29" containerID="3bec0e62246a8700e5ec85066fab5b59dc7b4d9c6faed38cabb0e8a5059162bc" exitCode=0 Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.296280 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88233d80-672d-47c3-9a26-70c22f748e29","Type":"ContainerDied","Data":"3bec0e62246a8700e5ec85066fab5b59dc7b4d9c6faed38cabb0e8a5059162bc"} Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.303610 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78wps" event={"ID":"79f551cd-d08d-4739-9041-4543f05fbb60","Type":"ContainerStarted","Data":"9552e98856386471a4652c417c2d11a2a782b4e19074de43beda6ae18f5f34d1"} Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.304017 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-78wps" Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.480670 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.506305 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-78wps" podStartSLOduration=22.010127904 podStartE2EDuration="50.506282621s" podCreationTimestamp="2025-12-08 19:40:04 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.515681505 +0000 UTC m=+1107.157882508" lastFinishedPulling="2025-12-08 19:40:53.011836222 +0000 UTC m=+1135.654037225" observedRunningTime="2025-12-08 19:40:54.379816212 +0000 UTC m=+1137.022017215" watchObservedRunningTime="2025-12-08 19:40:54.506282621 +0000 UTC m=+1137.148483624" Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.539621 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9kbjh"] Dec 08 19:40:54 crc kubenswrapper[4706]: I1208 19:40:54.539953 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" containerName="dnsmasq-dns" containerID="cri-o://3f95beea4e2b84fc7411a2303591cb075e6e10023ff74e91a2f2fd080473c7e0" gracePeriod=10 Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.329815 4706 generic.go:334] "Generic (PLEG): container finished" podID="b9f26f6c-226a-4026-bde7-10f4401f53e0" containerID="3f95beea4e2b84fc7411a2303591cb075e6e10023ff74e91a2f2fd080473c7e0" exitCode=0 Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.329992 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" event={"ID":"b9f26f6c-226a-4026-bde7-10f4401f53e0","Type":"ContainerDied","Data":"3f95beea4e2b84fc7411a2303591cb075e6e10023ff74e91a2f2fd080473c7e0"} Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.333288 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" event={"ID":"7811fdf0-d557-47cb-b700-4a072602f8c7","Type":"ContainerStarted","Data":"40d5cf0ac297ea46d43f146c9f1a9204572a9b3c928a6d5d47e948c039a20303"} Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.333738 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.336934 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"4d1845b4-bbc3-4661-83d4-72b972fa288a","Type":"ContainerStarted","Data":"d75305fcbe1477c094f08372c0870e2fc616c70937dc5dc3b7cae530894a6dac"} Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.337763 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.361952 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" podStartSLOduration=-9223371992.492846 podStartE2EDuration="44.361929381s" podCreationTimestamp="2025-12-08 19:40:12 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.813481981 +0000 UTC m=+1107.455682984" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:40:56.355773855 +0000 UTC m=+1138.997974858" watchObservedRunningTime="2025-12-08 19:40:56.361929381 +0000 UTC m=+1139.004130384" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.367721 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-bc75944f-m4nfc" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.399848 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-index-gateway-0" podStartSLOduration=-9223371992.454958 podStartE2EDuration="44.399818779s" podCreationTimestamp="2025-12-08 19:40:12 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.921904736 +0000 UTC m=+1107.564105739" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:40:56.383407502 +0000 UTC m=+1139.025608515" watchObservedRunningTime="2025-12-08 19:40:56.399818779 +0000 UTC m=+1139.042019782" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.570614 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.646340 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-dns-svc\") pod \"b9f26f6c-226a-4026-bde7-10f4401f53e0\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.646719 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmgtv\" (UniqueName: \"kubernetes.io/projected/b9f26f6c-226a-4026-bde7-10f4401f53e0-kube-api-access-nmgtv\") pod \"b9f26f6c-226a-4026-bde7-10f4401f53e0\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.646973 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-config\") pod \"b9f26f6c-226a-4026-bde7-10f4401f53e0\" (UID: \"b9f26f6c-226a-4026-bde7-10f4401f53e0\") " Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.658129 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f26f6c-226a-4026-bde7-10f4401f53e0-kube-api-access-nmgtv" (OuterVolumeSpecName: "kube-api-access-nmgtv") pod "b9f26f6c-226a-4026-bde7-10f4401f53e0" (UID: "b9f26f6c-226a-4026-bde7-10f4401f53e0"). InnerVolumeSpecName "kube-api-access-nmgtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.716737 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b9f26f6c-226a-4026-bde7-10f4401f53e0" (UID: "b9f26f6c-226a-4026-bde7-10f4401f53e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.728684 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-config" (OuterVolumeSpecName: "config") pod "b9f26f6c-226a-4026-bde7-10f4401f53e0" (UID: "b9f26f6c-226a-4026-bde7-10f4401f53e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.751175 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmgtv\" (UniqueName: \"kubernetes.io/projected/b9f26f6c-226a-4026-bde7-10f4401f53e0-kube-api-access-nmgtv\") on node \"crc\" DevicePath \"\"" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.751284 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:40:56 crc kubenswrapper[4706]: I1208 19:40:56.751299 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9f26f6c-226a-4026-bde7-10f4401f53e0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.364139 4706 generic.go:334] "Generic (PLEG): container finished" podID="13532f07-b125-4d67-bdb8-006b583bea40" containerID="c5e3902cfeb117ee9d279959c9f0d4dcd09139afb1f3bc7dac7186d6280903f5" exitCode=0 Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.364212 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"13532f07-b125-4d67-bdb8-006b583bea40","Type":"ContainerDied","Data":"c5e3902cfeb117ee9d279959c9f0d4dcd09139afb1f3bc7dac7186d6280903f5"} Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.368087 4706 generic.go:334] "Generic (PLEG): container finished" podID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerID="4d2558eb0a6a323f3921e28819b25ee1b7d270e9ab07768a3bfcedfd8570cdbd" exitCode=0 Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.368167 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ea8337d-cff3-414c-98f6-17b41653b4fb","Type":"ContainerDied","Data":"4d2558eb0a6a323f3921e28819b25ee1b7d270e9ab07768a3bfcedfd8570cdbd"} Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.371668 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.372427 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9kbjh" event={"ID":"b9f26f6c-226a-4026-bde7-10f4401f53e0","Type":"ContainerDied","Data":"510b78f609ebee0e06c1af21a7d7aebc9d74783f27b4e0f40f2137e8c575b1f5"} Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.372493 4706 scope.go:117] "RemoveContainer" containerID="3f95beea4e2b84fc7411a2303591cb075e6e10023ff74e91a2f2fd080473c7e0" Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.375017 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"908bed13-e8ef-4234-83fd-21b3f0ed218e","Type":"ContainerStarted","Data":"580be45a2242147e94e765fe95e7bb0fca947f855c446213b61b694f887e318f"} Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.440846 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=25.495674496 podStartE2EDuration="57.440823295s" podCreationTimestamp="2025-12-08 19:40:00 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.425454728 +0000 UTC m=+1107.067655731" lastFinishedPulling="2025-12-08 19:40:56.370603527 +0000 UTC m=+1139.012804530" observedRunningTime="2025-12-08 19:40:57.436830661 +0000 UTC m=+1140.079031664" watchObservedRunningTime="2025-12-08 19:40:57.440823295 +0000 UTC m=+1140.083024298" Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.495085 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9kbjh"] Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.505579 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9kbjh"] Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.625662 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" path="/var/lib/kubelet/pods/b9f26f6c-226a-4026-bde7-10f4401f53e0/volumes" Dec 08 19:40:57 crc kubenswrapper[4706]: I1208 19:40:57.920337 4706 scope.go:117] "RemoveContainer" containerID="0e36fae90b50af38a0c7268fc57d4f4210ccca86adffa06d94f3f473b19c6114" Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.394517 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ea8337d-cff3-414c-98f6-17b41653b4fb","Type":"ContainerStarted","Data":"3c8835b11dc1bed5c90dbdd9b422d6a651db9f230fb396efc8eaecee370eb7c3"} Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.396246 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.398710 4706 generic.go:334] "Generic (PLEG): container finished" podID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerID="b9240b7d3b9c39a64fdc169ca933af330b5f84a9bba65f7ebdcde2ee2fe3e519" exitCode=0 Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.398779 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1472b8cb-795a-4a57-95f8-184a7707e8ad","Type":"ContainerDied","Data":"b9240b7d3b9c39a64fdc169ca933af330b5f84a9bba65f7ebdcde2ee2fe3e519"} Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.403483 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"13532f07-b125-4d67-bdb8-006b583bea40","Type":"ContainerStarted","Data":"9e0cc4027aff3bbcd527e60b13fdaef3ffb93bb21470c8d4ef61924399798cd9"} Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.414453 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f498bda7-c379-4808-bd6c-8e5513c4322d","Type":"ContainerStarted","Data":"b57d9b3f707b7bb0235475d94d5dd1b00f5340a63c8dfd110863a72c871f4ccd"} Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.419074 4706 generic.go:334] "Generic (PLEG): container finished" podID="156710b7-7dbe-4bb7-9c69-ce726d4b7a4c" containerID="de2e8754471cd05408f23608f1b340aec92752980100888fff1b7851ac3d2124" exitCode=0 Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.419163 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c","Type":"ContainerDied","Data":"de2e8754471cd05408f23608f1b340aec92752980100888fff1b7851ac3d2124"} Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.433443 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88233d80-672d-47c3-9a26-70c22f748e29","Type":"ContainerStarted","Data":"db5994e5f3338aca62f4a5468269b253056b327a10267e7c8050607b6b88a795"} Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.446426 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.914893936 podStartE2EDuration="1m4.446396961s" podCreationTimestamp="2025-12-08 19:39:54 +0000 UTC" firstStartedPulling="2025-12-08 19:40:09.818375463 +0000 UTC m=+1092.460576466" lastFinishedPulling="2025-12-08 19:40:23.349878488 +0000 UTC m=+1105.992079491" observedRunningTime="2025-12-08 19:40:58.431388414 +0000 UTC m=+1141.073589447" watchObservedRunningTime="2025-12-08 19:40:58.446396961 +0000 UTC m=+1141.088597964" Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.532210 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=34.461211715 podStartE2EDuration="1m2.532178492s" podCreationTimestamp="2025-12-08 19:39:56 +0000 UTC" firstStartedPulling="2025-12-08 19:40:23.028680237 +0000 UTC m=+1105.670881240" lastFinishedPulling="2025-12-08 19:40:51.099647024 +0000 UTC m=+1133.741848017" observedRunningTime="2025-12-08 19:40:58.491536496 +0000 UTC m=+1141.133737499" watchObservedRunningTime="2025-12-08 19:40:58.532178492 +0000 UTC m=+1141.174379495" Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.558106 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=22.082664022 podStartE2EDuration="55.5580827s" podCreationTimestamp="2025-12-08 19:40:03 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.599396458 +0000 UTC m=+1107.241597461" lastFinishedPulling="2025-12-08 19:40:58.074815146 +0000 UTC m=+1140.717016139" observedRunningTime="2025-12-08 19:40:58.552061368 +0000 UTC m=+1141.194262371" watchObservedRunningTime="2025-12-08 19:40:58.5580827 +0000 UTC m=+1141.200283703" Dec 08 19:40:58 crc kubenswrapper[4706]: I1208 19:40:58.801646 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 08 19:40:59 crc kubenswrapper[4706]: I1208 19:40:59.495888 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1472b8cb-795a-4a57-95f8-184a7707e8ad","Type":"ContainerStarted","Data":"13404fedf045d4ea5efc3b98b3a810c9e6b8453164f7a0d6cc1b7e2ef389425e"} Dec 08 19:40:59 crc kubenswrapper[4706]: I1208 19:40:59.496864 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 08 19:40:59 crc kubenswrapper[4706]: I1208 19:40:59.511809 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"156710b7-7dbe-4bb7-9c69-ce726d4b7a4c","Type":"ContainerStarted","Data":"5abf2f751918127186701ec111afc8085c112916fcaccd2c8f9e25972b3081c2"} Dec 08 19:40:59 crc kubenswrapper[4706]: I1208 19:40:59.571005 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.258125692 podStartE2EDuration="1m6.570984125s" podCreationTimestamp="2025-12-08 19:39:53 +0000 UTC" firstStartedPulling="2025-12-08 19:39:55.976232445 +0000 UTC m=+1078.618433448" lastFinishedPulling="2025-12-08 19:40:23.289090878 +0000 UTC m=+1105.931291881" observedRunningTime="2025-12-08 19:40:59.561815035 +0000 UTC m=+1142.204016038" watchObservedRunningTime="2025-12-08 19:40:59.570984125 +0000 UTC m=+1142.213185128" Dec 08 19:40:59 crc kubenswrapper[4706]: I1208 19:40:59.801586 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.524650 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"40f5d973-ac12-469a-a1d9-aa871b7cc9a3","Type":"ContainerStarted","Data":"585bfe04b76007398b8eec44f3c224cd2b03b3ecb4e19a3580930a2e6057fc1c"} Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.550004 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371971.304792 podStartE2EDuration="1m5.549983596s" podCreationTimestamp="2025-12-08 19:39:55 +0000 UTC" firstStartedPulling="2025-12-08 19:40:23.78881853 +0000 UTC m=+1106.431019533" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:40:59.61366571 +0000 UTC m=+1142.255866713" watchObservedRunningTime="2025-12-08 19:41:00.549983596 +0000 UTC m=+1143.192184599" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.649045 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.835878 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=19.428006628 podStartE2EDuration="53.835850342s" podCreationTimestamp="2025-12-08 19:40:07 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.918933902 +0000 UTC m=+1107.561134905" lastFinishedPulling="2025-12-08 19:40:59.326777616 +0000 UTC m=+1141.968978619" observedRunningTime="2025-12-08 19:41:00.555320658 +0000 UTC m=+1143.197521661" watchObservedRunningTime="2025-12-08 19:41:00.835850342 +0000 UTC m=+1143.478051345" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.837915 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-x58g2"] Dec 08 19:41:00 crc kubenswrapper[4706]: E1208 19:41:00.838467 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" containerName="dnsmasq-dns" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.838497 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" containerName="dnsmasq-dns" Dec 08 19:41:00 crc kubenswrapper[4706]: E1208 19:41:00.838533 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" containerName="init" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.838543 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" containerName="init" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.838741 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f26f6c-226a-4026-bde7-10f4401f53e0" containerName="dnsmasq-dns" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.839916 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.889634 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-x58g2"] Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.950995 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.951078 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-config\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:00 crc kubenswrapper[4706]: I1208 19:41:00.951120 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8xnv\" (UniqueName: \"kubernetes.io/projected/d30df65d-b365-4424-970c-d83e7498bec7-kube-api-access-f8xnv\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.053103 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.053196 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-config\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.053238 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8xnv\" (UniqueName: \"kubernetes.io/projected/d30df65d-b365-4424-970c-d83e7498bec7-kube-api-access-f8xnv\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.054714 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-config\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.054758 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.106190 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8xnv\" (UniqueName: \"kubernetes.io/projected/d30df65d-b365-4424-970c-d83e7498bec7-kube-api-access-f8xnv\") pod \"dnsmasq-dns-7cb5889db5-x58g2\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.175972 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.867915 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.956567 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.963397 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.967238 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-vs6pc" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.967431 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.967470 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.967647 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 08 19:41:01 crc kubenswrapper[4706]: I1208 19:41:01.981213 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.072756 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.072827 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qrqz\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-kube-api-access-9qrqz\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.073295 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d81adfd7-689a-496a-b4bf-c35450b63ca9-cache\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.074705 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d81adfd7-689a-496a-b4bf-c35450b63ca9-lock\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.075287 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9ed548d1-6405-4df0-99f3-1800731a2a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ed548d1-6405-4df0-99f3-1800731a2a5b\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.178174 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d81adfd7-689a-496a-b4bf-c35450b63ca9-cache\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.178492 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d81adfd7-689a-496a-b4bf-c35450b63ca9-lock\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.178563 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9ed548d1-6405-4df0-99f3-1800731a2a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ed548d1-6405-4df0-99f3-1800731a2a5b\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.178641 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.178682 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qrqz\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-kube-api-access-9qrqz\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: E1208 19:41:02.178904 4706 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 19:41:02 crc kubenswrapper[4706]: E1208 19:41:02.178949 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.179016 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d81adfd7-689a-496a-b4bf-c35450b63ca9-cache\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: E1208 19:41:02.179040 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift podName:d81adfd7-689a-496a-b4bf-c35450b63ca9 nodeName:}" failed. No retries permitted until 2025-12-08 19:41:02.679009296 +0000 UTC m=+1145.321210479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift") pod "swift-storage-0" (UID: "d81adfd7-689a-496a-b4bf-c35450b63ca9") : configmap "swift-ring-files" not found Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.179109 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d81adfd7-689a-496a-b4bf-c35450b63ca9-lock\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.185150 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.185206 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9ed548d1-6405-4df0-99f3-1800731a2a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ed548d1-6405-4df0-99f3-1800731a2a5b\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7e0828e6578b395dc78021c3012cbc2763e473d2eb7471d3321b7cfab77ecd76/globalmount\"" pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.214786 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qrqz\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-kube-api-access-9qrqz\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.243444 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9ed548d1-6405-4df0-99f3-1800731a2a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ed548d1-6405-4df0-99f3-1800731a2a5b\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.527891 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-sht9s"] Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.531036 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.537821 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.540912 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.545781 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.580565 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88233d80-672d-47c3-9a26-70c22f748e29","Type":"ContainerStarted","Data":"289b3875c7e4260a02a828f1f523f9e8e25d8005d6513bbb6d751a180d517b6a"} Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.584517 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.591603 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-scripts\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.591789 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-combined-ca-bundle\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.592540 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-swiftconf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.592574 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-dispersionconf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.592660 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-ring-data-devices\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.592730 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5nzf\" (UniqueName: \"kubernetes.io/projected/652b348e-05c0-4cdf-826a-a9587a121921-kube-api-access-b5nzf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.592786 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/652b348e-05c0-4cdf-826a-a9587a121921-etc-swift\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.595921 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.638131 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-thlcc"] Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.640202 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.654017 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-sht9s"] Dec 08 19:41:02 crc kubenswrapper[4706]: E1208 19:41:02.661806 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-b5nzf ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-sht9s" podUID="652b348e-05c0-4cdf-826a-a9587a121921" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.664287 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-thlcc"] Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.664389 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-distributor-664b687b54-zfbc5" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.668711 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=27.830711625 podStartE2EDuration="1m1.668692051s" podCreationTimestamp="2025-12-08 19:40:01 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.231107767 +0000 UTC m=+1106.873308770" lastFinishedPulling="2025-12-08 19:40:58.069088193 +0000 UTC m=+1140.711289196" observedRunningTime="2025-12-08 19:41:02.628625671 +0000 UTC m=+1145.270826674" watchObservedRunningTime="2025-12-08 19:41:02.668692051 +0000 UTC m=+1145.310893054" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.700906 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-combined-ca-bundle\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701015 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-ring-data-devices\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701079 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701123 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-combined-ca-bundle\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701148 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-scripts\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701202 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-etc-swift\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701238 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-swiftconf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701284 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-dispersionconf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701314 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-dispersionconf\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701422 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-ring-data-devices\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701477 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhn7s\" (UniqueName: \"kubernetes.io/projected/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-kube-api-access-zhn7s\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701510 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5nzf\" (UniqueName: \"kubernetes.io/projected/652b348e-05c0-4cdf-826a-a9587a121921-kube-api-access-b5nzf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701546 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-swiftconf\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701563 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/652b348e-05c0-4cdf-826a-a9587a121921-etc-swift\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.701626 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-scripts\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: E1208 19:41:02.705149 4706 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 19:41:02 crc kubenswrapper[4706]: E1208 19:41:02.705193 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 19:41:02 crc kubenswrapper[4706]: E1208 19:41:02.705255 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift podName:d81adfd7-689a-496a-b4bf-c35450b63ca9 nodeName:}" failed. No retries permitted until 2025-12-08 19:41:03.705230531 +0000 UTC m=+1146.347431724 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift") pod "swift-storage-0" (UID: "d81adfd7-689a-496a-b4bf-c35450b63ca9") : configmap "swift-ring-files" not found Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.705754 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-sht9s"] Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.707944 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-scripts\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.708103 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/652b348e-05c0-4cdf-826a-a9587a121921-etc-swift\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.710305 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-ring-data-devices\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.716824 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-combined-ca-bundle\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.740925 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-dispersionconf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.744976 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-swiftconf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.745495 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5nzf\" (UniqueName: \"kubernetes.io/projected/652b348e-05c0-4cdf-826a-a9587a121921-kube-api-access-b5nzf\") pod \"swift-ring-rebalance-sht9s\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.803122 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-swiftconf\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.803295 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-ring-data-devices\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.803374 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-combined-ca-bundle\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.803405 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-scripts\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.803440 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-etc-swift\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.803471 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-dispersionconf\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.803526 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhn7s\" (UniqueName: \"kubernetes.io/projected/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-kube-api-access-zhn7s\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.805128 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-scripts\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.805169 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-ring-data-devices\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.805895 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-etc-swift\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.808579 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-dispersionconf\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.809910 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-combined-ca-bundle\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.809997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-swiftconf\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.831011 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhn7s\" (UniqueName: \"kubernetes.io/projected/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-kube-api-access-zhn7s\") pod \"swift-ring-rebalance-thlcc\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.962151 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-querier-5467947bf7-dxbfm" Dec 08 19:41:02 crc kubenswrapper[4706]: I1208 19:41:02.976532 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.109237 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.277833 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.325120 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.592757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.592913 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.608056 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.620037 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-ring-data-devices\") pod \"652b348e-05c0-4cdf-826a-a9587a121921\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.620153 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-dispersionconf\") pod \"652b348e-05c0-4cdf-826a-a9587a121921\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.620242 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-swiftconf\") pod \"652b348e-05c0-4cdf-826a-a9587a121921\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.620369 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5nzf\" (UniqueName: \"kubernetes.io/projected/652b348e-05c0-4cdf-826a-a9587a121921-kube-api-access-b5nzf\") pod \"652b348e-05c0-4cdf-826a-a9587a121921\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.620493 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-combined-ca-bundle\") pod \"652b348e-05c0-4cdf-826a-a9587a121921\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.620530 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-scripts\") pod \"652b348e-05c0-4cdf-826a-a9587a121921\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.620548 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/652b348e-05c0-4cdf-826a-a9587a121921-etc-swift\") pod \"652b348e-05c0-4cdf-826a-a9587a121921\" (UID: \"652b348e-05c0-4cdf-826a-a9587a121921\") " Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.625981 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/652b348e-05c0-4cdf-826a-a9587a121921-kube-api-access-b5nzf" (OuterVolumeSpecName: "kube-api-access-b5nzf") pod "652b348e-05c0-4cdf-826a-a9587a121921" (UID: "652b348e-05c0-4cdf-826a-a9587a121921"). InnerVolumeSpecName "kube-api-access-b5nzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.626161 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/652b348e-05c0-4cdf-826a-a9587a121921-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "652b348e-05c0-4cdf-826a-a9587a121921" (UID: "652b348e-05c0-4cdf-826a-a9587a121921"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.626763 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-scripts" (OuterVolumeSpecName: "scripts") pod "652b348e-05c0-4cdf-826a-a9587a121921" (UID: "652b348e-05c0-4cdf-826a-a9587a121921"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.626854 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "652b348e-05c0-4cdf-826a-a9587a121921" (UID: "652b348e-05c0-4cdf-826a-a9587a121921"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.633164 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "652b348e-05c0-4cdf-826a-a9587a121921" (UID: "652b348e-05c0-4cdf-826a-a9587a121921"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.641142 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "652b348e-05c0-4cdf-826a-a9587a121921" (UID: "652b348e-05c0-4cdf-826a-a9587a121921"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.643232 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "652b348e-05c0-4cdf-826a-a9587a121921" (UID: "652b348e-05c0-4cdf-826a-a9587a121921"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.723533 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.723989 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.724011 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:03 crc kubenswrapper[4706]: E1208 19:41:03.724033 4706 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 19:41:03 crc kubenswrapper[4706]: E1208 19:41:03.724076 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.724110 4706 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/652b348e-05c0-4cdf-826a-a9587a121921-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.724129 4706 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/652b348e-05c0-4cdf-826a-a9587a121921-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:03 crc kubenswrapper[4706]: E1208 19:41:03.724161 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift podName:d81adfd7-689a-496a-b4bf-c35450b63ca9 nodeName:}" failed. No retries permitted until 2025-12-08 19:41:05.724134768 +0000 UTC m=+1148.366335771 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift") pod "swift-storage-0" (UID: "d81adfd7-689a-496a-b4bf-c35450b63ca9") : configmap "swift-ring-files" not found Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.724204 4706 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.724223 4706 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/652b348e-05c0-4cdf-826a-a9587a121921-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:03 crc kubenswrapper[4706]: I1208 19:41:03.724238 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5nzf\" (UniqueName: \"kubernetes.io/projected/652b348e-05c0-4cdf-826a-a9587a121921-kube-api-access-b5nzf\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.338783 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.608673 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sht9s" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.667942 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-x58g2"] Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.700465 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vnr25"] Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.705561 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.721522 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.733513 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-sht9s"] Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.811440 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-sht9s"] Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.834200 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vnr25"] Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.860466 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.860554 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-config\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.860878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l5m5\" (UniqueName: \"kubernetes.io/projected/83042fbc-03d8-49ca-a870-a3dfd8343d72-kube-api-access-9l5m5\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.861025 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.887539 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-schr5"] Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.894932 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.895096 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.898940 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.903604 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-schr5"] Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.963302 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-config\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.963485 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l5m5\" (UniqueName: \"kubernetes.io/projected/83042fbc-03d8-49ca-a870-a3dfd8343d72-kube-api-access-9l5m5\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.963560 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.963650 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.964931 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.965771 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-config\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:04 crc kubenswrapper[4706]: I1208 19:41:04.966814 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.003025 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l5m5\" (UniqueName: \"kubernetes.io/projected/83042fbc-03d8-49ca-a870-a3dfd8343d72-kube-api-access-9l5m5\") pod \"dnsmasq-dns-6c89d5d749-vnr25\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.034438 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.069798 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a00cac1b-92f7-49a1-bbdc-97c15777b09a-ovs-rundir\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.069863 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a00cac1b-92f7-49a1-bbdc-97c15777b09a-config\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.069972 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00cac1b-92f7-49a1-bbdc-97c15777b09a-combined-ca-bundle\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.070007 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a00cac1b-92f7-49a1-bbdc-97c15777b09a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.070065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwrbm\" (UniqueName: \"kubernetes.io/projected/a00cac1b-92f7-49a1-bbdc-97c15777b09a-kube-api-access-mwrbm\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.070084 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a00cac1b-92f7-49a1-bbdc-97c15777b09a-ovn-rundir\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.152757 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vnr25"] Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.174578 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a00cac1b-92f7-49a1-bbdc-97c15777b09a-ovs-rundir\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.174667 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a00cac1b-92f7-49a1-bbdc-97c15777b09a-config\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.174744 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00cac1b-92f7-49a1-bbdc-97c15777b09a-combined-ca-bundle\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.174777 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a00cac1b-92f7-49a1-bbdc-97c15777b09a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.174850 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwrbm\" (UniqueName: \"kubernetes.io/projected/a00cac1b-92f7-49a1-bbdc-97c15777b09a-kube-api-access-mwrbm\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.174874 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a00cac1b-92f7-49a1-bbdc-97c15777b09a-ovn-rundir\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.175076 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a00cac1b-92f7-49a1-bbdc-97c15777b09a-ovn-rundir\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.175085 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a00cac1b-92f7-49a1-bbdc-97c15777b09a-ovs-rundir\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.176037 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a00cac1b-92f7-49a1-bbdc-97c15777b09a-config\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.189416 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00cac1b-92f7-49a1-bbdc-97c15777b09a-combined-ca-bundle\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.192016 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a00cac1b-92f7-49a1-bbdc-97c15777b09a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.211928 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-vs9wr"] Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.214082 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.218949 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.238253 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwrbm\" (UniqueName: \"kubernetes.io/projected/a00cac1b-92f7-49a1-bbdc-97c15777b09a-kube-api-access-mwrbm\") pod \"ovn-controller-metrics-schr5\" (UID: \"a00cac1b-92f7-49a1-bbdc-97c15777b09a\") " pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.258838 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-vs9wr"] Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.378335 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.378776 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-config\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.378859 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.378937 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-dns-svc\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.379030 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnlsq\" (UniqueName: \"kubernetes.io/projected/b17a1b80-f5a5-411a-a5bd-79aa80bed406-kube-api-access-rnlsq\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.485003 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.485103 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-dns-svc\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.485190 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnlsq\" (UniqueName: \"kubernetes.io/projected/b17a1b80-f5a5-411a-a5bd-79aa80bed406-kube-api-access-rnlsq\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.485220 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.485256 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-config\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.487153 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-dns-svc\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.491571 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-config\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.494083 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.495628 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.528206 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.530198 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.532529 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-schr5" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.534046 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnlsq\" (UniqueName: \"kubernetes.io/projected/b17a1b80-f5a5-411a-a5bd-79aa80bed406-kube-api-access-rnlsq\") pod \"dnsmasq-dns-698758b865-vs9wr\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.536808 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.537201 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.537502 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-wgw4b" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.537813 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.574706 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.619773 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.632211 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="652b348e-05c0-4cdf-826a-a9587a121921" path="/var/lib/kubelet/pods/652b348e-05c0-4cdf-826a-a9587a121921/volumes" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.689083 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.689171 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5592\" (UniqueName: \"kubernetes.io/projected/240f9d54-934b-4d65-976e-a0a5a9ca7908-kube-api-access-d5592\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.689203 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.689239 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/240f9d54-934b-4d65-976e-a0a5a9ca7908-scripts\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.689353 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/240f9d54-934b-4d65-976e-a0a5a9ca7908-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.689392 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.689423 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240f9d54-934b-4d65-976e-a0a5a9ca7908-config\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.791912 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/240f9d54-934b-4d65-976e-a0a5a9ca7908-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.792465 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/240f9d54-934b-4d65-976e-a0a5a9ca7908-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.793432 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.793533 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240f9d54-934b-4d65-976e-a0a5a9ca7908-config\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.793619 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.793775 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.793869 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5592\" (UniqueName: \"kubernetes.io/projected/240f9d54-934b-4d65-976e-a0a5a9ca7908-kube-api-access-d5592\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.793920 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.793990 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/240f9d54-934b-4d65-976e-a0a5a9ca7908-scripts\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.794622 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240f9d54-934b-4d65-976e-a0a5a9ca7908-config\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: E1208 19:41:05.794788 4706 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 19:41:05 crc kubenswrapper[4706]: E1208 19:41:05.794808 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 19:41:05 crc kubenswrapper[4706]: E1208 19:41:05.794861 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift podName:d81adfd7-689a-496a-b4bf-c35450b63ca9 nodeName:}" failed. No retries permitted until 2025-12-08 19:41:09.794843547 +0000 UTC m=+1152.437044550 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift") pod "swift-storage-0" (UID: "d81adfd7-689a-496a-b4bf-c35450b63ca9") : configmap "swift-ring-files" not found Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.795507 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/240f9d54-934b-4d65-976e-a0a5a9ca7908-scripts\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.800058 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.801732 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.803965 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/240f9d54-934b-4d65-976e-a0a5a9ca7908-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.816196 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5592\" (UniqueName: \"kubernetes.io/projected/240f9d54-934b-4d65-976e-a0a5a9ca7908-kube-api-access-d5592\") pod \"ovn-northd-0\" (UID: \"240f9d54-934b-4d65-976e-a0a5a9ca7908\") " pod="openstack/ovn-northd-0" Dec 08 19:41:05 crc kubenswrapper[4706]: I1208 19:41:05.889663 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 08 19:41:06 crc kubenswrapper[4706]: E1208 19:41:06.276894 4706 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:37720->38.102.83.113:43781: write tcp 38.102.83.113:37720->38.102.83.113:43781: write: broken pipe Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.470032 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vnr25"] Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.655224 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerStarted","Data":"5550e29689eec464a155d46fa108579fce71523aade86b30467ed46732e8ff81"} Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.663073 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" event={"ID":"83042fbc-03d8-49ca-a870-a3dfd8343d72","Type":"ContainerStarted","Data":"75f8acec686ea540e172c614d42799da5aecf4df663e2ab2cf71c2f843a8f2df"} Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.697377 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-thlcc"] Dec 08 19:41:06 crc kubenswrapper[4706]: W1208 19:41:06.704399 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7d7a844_b2ae_477f_b8c0_d14f8ec87d3c.slice/crio-d5d871c568e64473eb7f698d05003e39532b2043983a99a477558022e051f3cb WatchSource:0}: Error finding container d5d871c568e64473eb7f698d05003e39532b2043983a99a477558022e051f3cb: Status 404 returned error can't find the container with id d5d871c568e64473eb7f698d05003e39532b2043983a99a477558022e051f3cb Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.725416 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-vs9wr"] Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.737723 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-schr5"] Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.741195 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.741276 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 08 19:41:06 crc kubenswrapper[4706]: W1208 19:41:06.811210 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb17a1b80_f5a5_411a_a5bd_79aa80bed406.slice/crio-bb5bd212bf1b0aa5921edac41b16fb83f0c136cfec36118fa768077684cd2081 WatchSource:0}: Error finding container bb5bd212bf1b0aa5921edac41b16fb83f0c136cfec36118fa768077684cd2081: Status 404 returned error can't find the container with id bb5bd212bf1b0aa5921edac41b16fb83f0c136cfec36118fa768077684cd2081 Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.901049 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-x58g2"] Dec 08 19:41:06 crc kubenswrapper[4706]: W1208 19:41:06.913678 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd30df65d_b365_4424_970c_d83e7498bec7.slice/crio-1bed4018c2438569caf094df3d41c4c7c1b2921eb2d47fba8f396865107420e7 WatchSource:0}: Error finding container 1bed4018c2438569caf094df3d41c4c7c1b2921eb2d47fba8f396865107420e7: Status 404 returned error can't find the container with id 1bed4018c2438569caf094df3d41c4c7c1b2921eb2d47fba8f396865107420e7 Dec 08 19:41:06 crc kubenswrapper[4706]: I1208 19:41:06.956498 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 08 19:41:07 crc kubenswrapper[4706]: W1208 19:41:07.022637 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod240f9d54_934b_4d65_976e_a0a5a9ca7908.slice/crio-76046683cf297135c54b45a42a656b167c80de6a33224a2042c5ebecba52fd0b WatchSource:0}: Error finding container 76046683cf297135c54b45a42a656b167c80de6a33224a2042c5ebecba52fd0b: Status 404 returned error can't find the container with id 76046683cf297135c54b45a42a656b167c80de6a33224a2042c5ebecba52fd0b Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.684373 4706 generic.go:334] "Generic (PLEG): container finished" podID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerID="a8533ddb4be48988ddd60b6dd0daa9ea77760010e7a4f004a6979806bb7e0ec6" exitCode=0 Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.684499 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-vs9wr" event={"ID":"b17a1b80-f5a5-411a-a5bd-79aa80bed406","Type":"ContainerDied","Data":"a8533ddb4be48988ddd60b6dd0daa9ea77760010e7a4f004a6979806bb7e0ec6"} Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.690998 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-vs9wr" event={"ID":"b17a1b80-f5a5-411a-a5bd-79aa80bed406","Type":"ContainerStarted","Data":"bb5bd212bf1b0aa5921edac41b16fb83f0c136cfec36118fa768077684cd2081"} Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.691189 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-thlcc" event={"ID":"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c","Type":"ContainerStarted","Data":"d5d871c568e64473eb7f698d05003e39532b2043983a99a477558022e051f3cb"} Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.693184 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"240f9d54-934b-4d65-976e-a0a5a9ca7908","Type":"ContainerStarted","Data":"76046683cf297135c54b45a42a656b167c80de6a33224a2042c5ebecba52fd0b"} Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.695788 4706 generic.go:334] "Generic (PLEG): container finished" podID="83042fbc-03d8-49ca-a870-a3dfd8343d72" containerID="b47c13f3929920247b098fff5c10c1c9a57738de34d7d45050853f1c3bd4cbd0" exitCode=0 Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.696026 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" event={"ID":"83042fbc-03d8-49ca-a870-a3dfd8343d72","Type":"ContainerDied","Data":"b47c13f3929920247b098fff5c10c1c9a57738de34d7d45050853f1c3bd4cbd0"} Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.713605 4706 generic.go:334] "Generic (PLEG): container finished" podID="d30df65d-b365-4424-970c-d83e7498bec7" containerID="9f1b14e09009d9417351c348952cffe4a7cc295dd68da7fb6dda63740ae424f7" exitCode=0 Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.714218 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" event={"ID":"d30df65d-b365-4424-970c-d83e7498bec7","Type":"ContainerDied","Data":"9f1b14e09009d9417351c348952cffe4a7cc295dd68da7fb6dda63740ae424f7"} Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.714363 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" event={"ID":"d30df65d-b365-4424-970c-d83e7498bec7","Type":"ContainerStarted","Data":"1bed4018c2438569caf094df3d41c4c7c1b2921eb2d47fba8f396865107420e7"} Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.727832 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-schr5" event={"ID":"a00cac1b-92f7-49a1-bbdc-97c15777b09a","Type":"ContainerStarted","Data":"738327c627ebf2d3dff346fa6bd9ce9fe8254936bbd12dd14b133c43d35d83b1"} Dec 08 19:41:07 crc kubenswrapper[4706]: I1208 19:41:07.728306 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-schr5" event={"ID":"a00cac1b-92f7-49a1-bbdc-97c15777b09a","Type":"ContainerStarted","Data":"55217e0b2aa1adca037d48a8954c3749ceffc140ab08f13ebb64df42c755a930"} Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.016675 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-schr5" podStartSLOduration=4.016641137 podStartE2EDuration="4.016641137s" podCreationTimestamp="2025-12-08 19:41:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:07.992395297 +0000 UTC m=+1150.634596310" watchObservedRunningTime="2025-12-08 19:41:08.016641137 +0000 UTC m=+1150.658842140" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.258809 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.258865 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.428094 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.712908 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.718614 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.766709 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-vs9wr" event={"ID":"b17a1b80-f5a5-411a-a5bd-79aa80bed406","Type":"ContainerStarted","Data":"c1b660ef7677eefaf41a2b192223ca816a3de8b32ee1f06fe79f22e5b7435ee3"} Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.766884 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.772379 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.773042 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vnr25" event={"ID":"83042fbc-03d8-49ca-a870-a3dfd8343d72","Type":"ContainerDied","Data":"75f8acec686ea540e172c614d42799da5aecf4df663e2ab2cf71c2f843a8f2df"} Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.773097 4706 scope.go:117] "RemoveContainer" containerID="b47c13f3929920247b098fff5c10c1c9a57738de34d7d45050853f1c3bd4cbd0" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.778482 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-dns-svc\") pod \"d30df65d-b365-4424-970c-d83e7498bec7\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.778566 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l5m5\" (UniqueName: \"kubernetes.io/projected/83042fbc-03d8-49ca-a870-a3dfd8343d72-kube-api-access-9l5m5\") pod \"83042fbc-03d8-49ca-a870-a3dfd8343d72\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.779242 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-config\") pod \"d30df65d-b365-4424-970c-d83e7498bec7\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.779338 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8xnv\" (UniqueName: \"kubernetes.io/projected/d30df65d-b365-4424-970c-d83e7498bec7-kube-api-access-f8xnv\") pod \"d30df65d-b365-4424-970c-d83e7498bec7\" (UID: \"d30df65d-b365-4424-970c-d83e7498bec7\") " Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.779470 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-ovsdbserver-sb\") pod \"83042fbc-03d8-49ca-a870-a3dfd8343d72\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.779500 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-dns-svc\") pod \"83042fbc-03d8-49ca-a870-a3dfd8343d72\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.779596 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-config\") pod \"83042fbc-03d8-49ca-a870-a3dfd8343d72\" (UID: \"83042fbc-03d8-49ca-a870-a3dfd8343d72\") " Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.785097 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" event={"ID":"d30df65d-b365-4424-970c-d83e7498bec7","Type":"ContainerDied","Data":"1bed4018c2438569caf094df3d41c4c7c1b2921eb2d47fba8f396865107420e7"} Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.785601 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-x58g2" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.788171 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83042fbc-03d8-49ca-a870-a3dfd8343d72-kube-api-access-9l5m5" (OuterVolumeSpecName: "kube-api-access-9l5m5") pod "83042fbc-03d8-49ca-a870-a3dfd8343d72" (UID: "83042fbc-03d8-49ca-a870-a3dfd8343d72"). InnerVolumeSpecName "kube-api-access-9l5m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.788232 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d30df65d-b365-4424-970c-d83e7498bec7-kube-api-access-f8xnv" (OuterVolumeSpecName: "kube-api-access-f8xnv") pod "d30df65d-b365-4424-970c-d83e7498bec7" (UID: "d30df65d-b365-4424-970c-d83e7498bec7"). InnerVolumeSpecName "kube-api-access-f8xnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.812824 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-vs9wr" podStartSLOduration=3.812803424 podStartE2EDuration="3.812803424s" podCreationTimestamp="2025-12-08 19:41:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:08.806503184 +0000 UTC m=+1151.448704187" watchObservedRunningTime="2025-12-08 19:41:08.812803424 +0000 UTC m=+1151.455004427" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.884245 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l5m5\" (UniqueName: \"kubernetes.io/projected/83042fbc-03d8-49ca-a870-a3dfd8343d72-kube-api-access-9l5m5\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.884607 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8xnv\" (UniqueName: \"kubernetes.io/projected/d30df65d-b365-4424-970c-d83e7498bec7-kube-api-access-f8xnv\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:08 crc kubenswrapper[4706]: I1208 19:41:08.910864 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.016072 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-config" (OuterVolumeSpecName: "config") pod "d30df65d-b365-4424-970c-d83e7498bec7" (UID: "d30df65d-b365-4424-970c-d83e7498bec7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.044387 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.088707 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.207184 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d30df65d-b365-4424-970c-d83e7498bec7" (UID: "d30df65d-b365-4424-970c-d83e7498bec7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.215177 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "83042fbc-03d8-49ca-a870-a3dfd8343d72" (UID: "83042fbc-03d8-49ca-a870-a3dfd8343d72"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.216371 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-config" (OuterVolumeSpecName: "config") pod "83042fbc-03d8-49ca-a870-a3dfd8343d72" (UID: "83042fbc-03d8-49ca-a870-a3dfd8343d72"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.219885 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "83042fbc-03d8-49ca-a870-a3dfd8343d72" (UID: "83042fbc-03d8-49ca-a870-a3dfd8343d72"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.230245 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.293998 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.294050 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.294062 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83042fbc-03d8-49ca-a870-a3dfd8343d72-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.294075 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d30df65d-b365-4424-970c-d83e7498bec7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.472369 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vnr25"] Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.491251 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vnr25"] Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.520996 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-x58g2"] Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.533632 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-x58g2"] Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.630246 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83042fbc-03d8-49ca-a870-a3dfd8343d72" path="/var/lib/kubelet/pods/83042fbc-03d8-49ca-a870-a3dfd8343d72/volumes" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.630917 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d30df65d-b365-4424-970c-d83e7498bec7" path="/var/lib/kubelet/pods/d30df65d-b365-4424-970c-d83e7498bec7/volumes" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.685442 4706 scope.go:117] "RemoveContainer" containerID="9f1b14e09009d9417351c348952cffe4a7cc295dd68da7fb6dda63740ae424f7" Dec 08 19:41:09 crc kubenswrapper[4706]: I1208 19:41:09.805176 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:09 crc kubenswrapper[4706]: E1208 19:41:09.805557 4706 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 19:41:09 crc kubenswrapper[4706]: E1208 19:41:09.805677 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 19:41:09 crc kubenswrapper[4706]: E1208 19:41:09.805731 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift podName:d81adfd7-689a-496a-b4bf-c35450b63ca9 nodeName:}" failed. No retries permitted until 2025-12-08 19:41:17.80571174 +0000 UTC m=+1160.447912743 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift") pod "swift-storage-0" (UID: "d81adfd7-689a-496a-b4bf-c35450b63ca9") : configmap "swift-ring-files" not found Dec 08 19:41:10 crc kubenswrapper[4706]: I1208 19:41:10.654637 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 08 19:41:10 crc kubenswrapper[4706]: I1208 19:41:10.817328 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerStarted","Data":"42582ba53546784a04ba0ac07e86af8166caf6957c42a3d67bd4cfa7e3bad73f"} Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.767942 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c9ab-account-create-update-b72xz"] Dec 08 19:41:13 crc kubenswrapper[4706]: E1208 19:41:13.775222 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83042fbc-03d8-49ca-a870-a3dfd8343d72" containerName="init" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.775245 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="83042fbc-03d8-49ca-a870-a3dfd8343d72" containerName="init" Dec 08 19:41:13 crc kubenswrapper[4706]: E1208 19:41:13.775282 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30df65d-b365-4424-970c-d83e7498bec7" containerName="init" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.775288 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30df65d-b365-4424-970c-d83e7498bec7" containerName="init" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.775494 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d30df65d-b365-4424-970c-d83e7498bec7" containerName="init" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.775520 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="83042fbc-03d8-49ca-a870-a3dfd8343d72" containerName="init" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.776801 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.788679 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.805494 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-operator-scripts\") pod \"glance-c9ab-account-create-update-b72xz\" (UID: \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\") " pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.805721 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8grc\" (UniqueName: \"kubernetes.io/projected/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-kube-api-access-s8grc\") pod \"glance-c9ab-account-create-update-b72xz\" (UID: \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\") " pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.824598 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c9ab-account-create-update-b72xz"] Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.843337 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-pvxvl"] Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.845167 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.870779 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-thlcc" event={"ID":"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c","Type":"ContainerStarted","Data":"a3dc3350128b2066c05ed3e4109abe891d8d32b47f84bc2a2eb2876628f0c10d"} Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.877756 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pvxvl"] Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.879540 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"240f9d54-934b-4d65-976e-a0a5a9ca7908","Type":"ContainerStarted","Data":"841985f139574c12c0f60dd69bed6314f1f55d10b2eabd4e1b8243932ff10bed"} Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.898824 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="9a912025-8f8e-4d61-861e-c933dafe4660" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.908114 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-operator-scripts\") pod \"glance-c9ab-account-create-update-b72xz\" (UID: \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\") " pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.908209 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghc2w\" (UniqueName: \"kubernetes.io/projected/a36b48c2-0389-416b-a142-b28dc1ea76da-kube-api-access-ghc2w\") pod \"glance-db-create-pvxvl\" (UID: \"a36b48c2-0389-416b-a142-b28dc1ea76da\") " pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.908244 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8grc\" (UniqueName: \"kubernetes.io/projected/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-kube-api-access-s8grc\") pod \"glance-c9ab-account-create-update-b72xz\" (UID: \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\") " pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.908328 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a36b48c2-0389-416b-a142-b28dc1ea76da-operator-scripts\") pod \"glance-db-create-pvxvl\" (UID: \"a36b48c2-0389-416b-a142-b28dc1ea76da\") " pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.910214 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-operator-scripts\") pod \"glance-c9ab-account-create-update-b72xz\" (UID: \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\") " pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:13 crc kubenswrapper[4706]: I1208 19:41:13.930050 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8grc\" (UniqueName: \"kubernetes.io/projected/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-kube-api-access-s8grc\") pod \"glance-c9ab-account-create-update-b72xz\" (UID: \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\") " pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.010865 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a36b48c2-0389-416b-a142-b28dc1ea76da-operator-scripts\") pod \"glance-db-create-pvxvl\" (UID: \"a36b48c2-0389-416b-a142-b28dc1ea76da\") " pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.011060 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghc2w\" (UniqueName: \"kubernetes.io/projected/a36b48c2-0389-416b-a142-b28dc1ea76da-kube-api-access-ghc2w\") pod \"glance-db-create-pvxvl\" (UID: \"a36b48c2-0389-416b-a142-b28dc1ea76da\") " pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.011706 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a36b48c2-0389-416b-a142-b28dc1ea76da-operator-scripts\") pod \"glance-db-create-pvxvl\" (UID: \"a36b48c2-0389-416b-a142-b28dc1ea76da\") " pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.032391 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghc2w\" (UniqueName: \"kubernetes.io/projected/a36b48c2-0389-416b-a142-b28dc1ea76da-kube-api-access-ghc2w\") pod \"glance-db-create-pvxvl\" (UID: \"a36b48c2-0389-416b-a142-b28dc1ea76da\") " pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.061673 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.092007 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-thlcc" podStartSLOduration=5.346385212 podStartE2EDuration="12.091978901s" podCreationTimestamp="2025-12-08 19:41:02 +0000 UTC" firstStartedPulling="2025-12-08 19:41:06.707173111 +0000 UTC m=+1149.349374114" lastFinishedPulling="2025-12-08 19:41:13.4527668 +0000 UTC m=+1156.094967803" observedRunningTime="2025-12-08 19:41:13.893425201 +0000 UTC m=+1156.535626214" watchObservedRunningTime="2025-12-08 19:41:14.091978901 +0000 UTC m=+1156.734179904" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.124872 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.178241 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.222135 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.773795 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c9ab-account-create-update-b72xz"] Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.900408 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c9ab-account-create-update-b72xz" event={"ID":"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56","Type":"ContainerStarted","Data":"8b2daf4c0fa100aa65f11cd97e7fd79a4b53132309f2635206cdfa8e89dfb1a2"} Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.904248 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"240f9d54-934b-4d65-976e-a0a5a9ca7908","Type":"ContainerStarted","Data":"3cb80c8bf245d9bee21da4a77e54d27f0c72f8059bf9ef43efcaa9e77d1a2df1"} Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.904368 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.943341 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pvxvl"] Dec 08 19:41:14 crc kubenswrapper[4706]: I1208 19:41:14.947898 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.529641636 podStartE2EDuration="9.947875889s" podCreationTimestamp="2025-12-08 19:41:05 +0000 UTC" firstStartedPulling="2025-12-08 19:41:07.026337974 +0000 UTC m=+1149.668538977" lastFinishedPulling="2025-12-08 19:41:13.444572227 +0000 UTC m=+1156.086773230" observedRunningTime="2025-12-08 19:41:14.937395711 +0000 UTC m=+1157.579596714" watchObservedRunningTime="2025-12-08 19:41:14.947875889 +0000 UTC m=+1157.590076892" Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.191584 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.626700 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.697765 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.838358 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lqbnv"] Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.838690 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" containerName="dnsmasq-dns" containerID="cri-o://ae486bf81a5b12911f8587408d3868244a134566a5e315c09dac14d9c7bff8e9" gracePeriod=10 Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.926662 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec9a-account-create-update-bfpsl"] Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.928610 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.934797 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.960375 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c9ab-account-create-update-b72xz" event={"ID":"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56","Type":"ContainerStarted","Data":"03710859b1f1f8aede28d1c0570fe1e5ee1b7eb978dd4f839d9e220d609766bc"} Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.996435 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pvxvl" event={"ID":"a36b48c2-0389-416b-a142-b28dc1ea76da","Type":"ContainerStarted","Data":"286bb523070edb6a5131de79c1e7d90a4a41362b076e9aa5d5b6b2a0e851e984"} Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.996497 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pvxvl" event={"ID":"a36b48c2-0389-416b-a142-b28dc1ea76da","Type":"ContainerStarted","Data":"6f9e6eff7aa050660d3c8b7cf8cf1732d2a1b09d9bccf03e02b4360b88b7a2c7"} Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.997621 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r28zg\" (UniqueName: \"kubernetes.io/projected/f23ba90a-3542-4241-9e27-63a9bcd2baee-kube-api-access-r28zg\") pod \"cinder-ec9a-account-create-update-bfpsl\" (UID: \"f23ba90a-3542-4241-9e27-63a9bcd2baee\") " pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:15 crc kubenswrapper[4706]: I1208 19:41:15.997676 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f23ba90a-3542-4241-9e27-63a9bcd2baee-operator-scripts\") pod \"cinder-ec9a-account-create-update-bfpsl\" (UID: \"f23ba90a-3542-4241-9e27-63a9bcd2baee\") " pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.013816 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-szsrk"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.015815 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.068355 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec9a-account-create-update-bfpsl"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.093230 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-szsrk"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.107867 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r28zg\" (UniqueName: \"kubernetes.io/projected/f23ba90a-3542-4241-9e27-63a9bcd2baee-kube-api-access-r28zg\") pod \"cinder-ec9a-account-create-update-bfpsl\" (UID: \"f23ba90a-3542-4241-9e27-63a9bcd2baee\") " pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.107933 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f23ba90a-3542-4241-9e27-63a9bcd2baee-operator-scripts\") pod \"cinder-ec9a-account-create-update-bfpsl\" (UID: \"f23ba90a-3542-4241-9e27-63a9bcd2baee\") " pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.107957 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/091f4a40-69b6-4c09-8251-d5aca75a420f-operator-scripts\") pod \"cinder-db-create-szsrk\" (UID: \"091f4a40-69b6-4c09-8251-d5aca75a420f\") " pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.108040 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmwzd\" (UniqueName: \"kubernetes.io/projected/091f4a40-69b6-4c09-8251-d5aca75a420f-kube-api-access-rmwzd\") pod \"cinder-db-create-szsrk\" (UID: \"091f4a40-69b6-4c09-8251-d5aca75a420f\") " pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.112173 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f23ba90a-3542-4241-9e27-63a9bcd2baee-operator-scripts\") pod \"cinder-ec9a-account-create-update-bfpsl\" (UID: \"f23ba90a-3542-4241-9e27-63a9bcd2baee\") " pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.141732 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-pjffv"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.151452 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.169048 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r28zg\" (UniqueName: \"kubernetes.io/projected/f23ba90a-3542-4241-9e27-63a9bcd2baee-kube-api-access-r28zg\") pod \"cinder-ec9a-account-create-update-bfpsl\" (UID: \"f23ba90a-3542-4241-9e27-63a9bcd2baee\") " pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.186226 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-pjffv"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.188429 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-c9ab-account-create-update-b72xz" podStartSLOduration=3.188413862 podStartE2EDuration="3.188413862s" podCreationTimestamp="2025-12-08 19:41:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:16.03970649 +0000 UTC m=+1158.681907493" watchObservedRunningTime="2025-12-08 19:41:16.188413862 +0000 UTC m=+1158.830614865" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.204857 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-pvxvl" podStartSLOduration=3.20482706 podStartE2EDuration="3.20482706s" podCreationTimestamp="2025-12-08 19:41:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:16.068103759 +0000 UTC m=+1158.710304782" watchObservedRunningTime="2025-12-08 19:41:16.20482706 +0000 UTC m=+1158.847028063" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.210593 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb919dfd-81a4-42b1-8e30-4788113753d2-operator-scripts\") pod \"barbican-db-create-pjffv\" (UID: \"bb919dfd-81a4-42b1-8e30-4788113753d2\") " pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.210975 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmwzd\" (UniqueName: \"kubernetes.io/projected/091f4a40-69b6-4c09-8251-d5aca75a420f-kube-api-access-rmwzd\") pod \"cinder-db-create-szsrk\" (UID: \"091f4a40-69b6-4c09-8251-d5aca75a420f\") " pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.211211 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdg99\" (UniqueName: \"kubernetes.io/projected/bb919dfd-81a4-42b1-8e30-4788113753d2-kube-api-access-cdg99\") pod \"barbican-db-create-pjffv\" (UID: \"bb919dfd-81a4-42b1-8e30-4788113753d2\") " pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.211412 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/091f4a40-69b6-4c09-8251-d5aca75a420f-operator-scripts\") pod \"cinder-db-create-szsrk\" (UID: \"091f4a40-69b6-4c09-8251-d5aca75a420f\") " pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.215066 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/091f4a40-69b6-4c09-8251-d5aca75a420f-operator-scripts\") pod \"cinder-db-create-szsrk\" (UID: \"091f4a40-69b6-4c09-8251-d5aca75a420f\") " pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.253008 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b7de-account-create-update-ntz9l"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.254888 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.262352 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.273819 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.302184 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmwzd\" (UniqueName: \"kubernetes.io/projected/091f4a40-69b6-4c09-8251-d5aca75a420f-kube-api-access-rmwzd\") pod \"cinder-db-create-szsrk\" (UID: \"091f4a40-69b6-4c09-8251-d5aca75a420f\") " pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.315684 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h475g\" (UniqueName: \"kubernetes.io/projected/5a7451c2-8247-406f-9d8f-eb5690a16555-kube-api-access-h475g\") pod \"barbican-b7de-account-create-update-ntz9l\" (UID: \"5a7451c2-8247-406f-9d8f-eb5690a16555\") " pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.315776 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdg99\" (UniqueName: \"kubernetes.io/projected/bb919dfd-81a4-42b1-8e30-4788113753d2-kube-api-access-cdg99\") pod \"barbican-db-create-pjffv\" (UID: \"bb919dfd-81a4-42b1-8e30-4788113753d2\") " pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.315853 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7451c2-8247-406f-9d8f-eb5690a16555-operator-scripts\") pod \"barbican-b7de-account-create-update-ntz9l\" (UID: \"5a7451c2-8247-406f-9d8f-eb5690a16555\") " pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.315950 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb919dfd-81a4-42b1-8e30-4788113753d2-operator-scripts\") pod \"barbican-db-create-pjffv\" (UID: \"bb919dfd-81a4-42b1-8e30-4788113753d2\") " pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.344758 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb919dfd-81a4-42b1-8e30-4788113753d2-operator-scripts\") pod \"barbican-db-create-pjffv\" (UID: \"bb919dfd-81a4-42b1-8e30-4788113753d2\") " pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.361654 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b7de-account-create-update-ntz9l"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.365939 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.398178 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdg99\" (UniqueName: \"kubernetes.io/projected/bb919dfd-81a4-42b1-8e30-4788113753d2-kube-api-access-cdg99\") pod \"barbican-db-create-pjffv\" (UID: \"bb919dfd-81a4-42b1-8e30-4788113753d2\") " pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.418307 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h475g\" (UniqueName: \"kubernetes.io/projected/5a7451c2-8247-406f-9d8f-eb5690a16555-kube-api-access-h475g\") pod \"barbican-b7de-account-create-update-ntz9l\" (UID: \"5a7451c2-8247-406f-9d8f-eb5690a16555\") " pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.418472 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7451c2-8247-406f-9d8f-eb5690a16555-operator-scripts\") pod \"barbican-b7de-account-create-update-ntz9l\" (UID: \"5a7451c2-8247-406f-9d8f-eb5690a16555\") " pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.419589 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7451c2-8247-406f-9d8f-eb5690a16555-operator-scripts\") pod \"barbican-b7de-account-create-update-ntz9l\" (UID: \"5a7451c2-8247-406f-9d8f-eb5690a16555\") " pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.452009 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-create-6lf8v"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.453626 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dfae-account-create-update-dhrlq"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.453796 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.459837 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.468557 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.480765 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h475g\" (UniqueName: \"kubernetes.io/projected/5a7451c2-8247-406f-9d8f-eb5690a16555-kube-api-access-h475g\") pod \"barbican-b7de-account-create-update-ntz9l\" (UID: \"5a7451c2-8247-406f-9d8f-eb5690a16555\") " pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.480847 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-6lf8v"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.493386 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dfae-account-create-update-dhrlq"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.515517 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-xlsrb"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.517730 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.523646 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nlzj\" (UniqueName: \"kubernetes.io/projected/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-kube-api-access-5nlzj\") pod \"cloudkitty-db-create-6lf8v\" (UID: \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\") " pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.523752 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53d89be3-c6e5-4f3e-b788-4811b47ba332-operator-scripts\") pod \"neutron-dfae-account-create-update-dhrlq\" (UID: \"53d89be3-c6e5-4f3e-b788-4811b47ba332\") " pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.526512 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trbwj\" (UniqueName: \"kubernetes.io/projected/53d89be3-c6e5-4f3e-b788-4811b47ba332-kube-api-access-trbwj\") pod \"neutron-dfae-account-create-update-dhrlq\" (UID: \"53d89be3-c6e5-4f3e-b788-4811b47ba332\") " pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.529576 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-3e4b-account-create-update-ghzkt"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.530127 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-operator-scripts\") pod \"cloudkitty-db-create-6lf8v\" (UID: \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\") " pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.531453 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.533714 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-db-secret" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.538321 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.552193 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xlsrb"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.627190 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-3e4b-account-create-update-ghzkt"] Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.640372 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsccl\" (UniqueName: \"kubernetes.io/projected/a2583ff3-e358-446c-9660-db7074583d8a-kube-api-access-jsccl\") pod \"cloudkitty-3e4b-account-create-update-ghzkt\" (UID: \"a2583ff3-e358-446c-9660-db7074583d8a\") " pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.641449 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nlzj\" (UniqueName: \"kubernetes.io/projected/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-kube-api-access-5nlzj\") pod \"cloudkitty-db-create-6lf8v\" (UID: \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\") " pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.641604 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53d89be3-c6e5-4f3e-b788-4811b47ba332-operator-scripts\") pod \"neutron-dfae-account-create-update-dhrlq\" (UID: \"53d89be3-c6e5-4f3e-b788-4811b47ba332\") " pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.641721 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trbwj\" (UniqueName: \"kubernetes.io/projected/53d89be3-c6e5-4f3e-b788-4811b47ba332-kube-api-access-trbwj\") pod \"neutron-dfae-account-create-update-dhrlq\" (UID: \"53d89be3-c6e5-4f3e-b788-4811b47ba332\") " pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.641793 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6507d8ce-f566-46d6-b55d-abbaccc71857-operator-scripts\") pod \"neutron-db-create-xlsrb\" (UID: \"6507d8ce-f566-46d6-b55d-abbaccc71857\") " pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.646867 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.657810 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53d89be3-c6e5-4f3e-b788-4811b47ba332-operator-scripts\") pod \"neutron-dfae-account-create-update-dhrlq\" (UID: \"53d89be3-c6e5-4f3e-b788-4811b47ba332\") " pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.658633 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t2n6\" (UniqueName: \"kubernetes.io/projected/6507d8ce-f566-46d6-b55d-abbaccc71857-kube-api-access-6t2n6\") pod \"neutron-db-create-xlsrb\" (UID: \"6507d8ce-f566-46d6-b55d-abbaccc71857\") " pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.658805 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-operator-scripts\") pod \"cloudkitty-db-create-6lf8v\" (UID: \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\") " pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.658847 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2583ff3-e358-446c-9660-db7074583d8a-operator-scripts\") pod \"cloudkitty-3e4b-account-create-update-ghzkt\" (UID: \"a2583ff3-e358-446c-9660-db7074583d8a\") " pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.660485 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-operator-scripts\") pod \"cloudkitty-db-create-6lf8v\" (UID: \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\") " pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.682700 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nlzj\" (UniqueName: \"kubernetes.io/projected/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-kube-api-access-5nlzj\") pod \"cloudkitty-db-create-6lf8v\" (UID: \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\") " pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.683538 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trbwj\" (UniqueName: \"kubernetes.io/projected/53d89be3-c6e5-4f3e-b788-4811b47ba332-kube-api-access-trbwj\") pod \"neutron-dfae-account-create-update-dhrlq\" (UID: \"53d89be3-c6e5-4f3e-b788-4811b47ba332\") " pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.763121 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6507d8ce-f566-46d6-b55d-abbaccc71857-operator-scripts\") pod \"neutron-db-create-xlsrb\" (UID: \"6507d8ce-f566-46d6-b55d-abbaccc71857\") " pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.763247 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t2n6\" (UniqueName: \"kubernetes.io/projected/6507d8ce-f566-46d6-b55d-abbaccc71857-kube-api-access-6t2n6\") pod \"neutron-db-create-xlsrb\" (UID: \"6507d8ce-f566-46d6-b55d-abbaccc71857\") " pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.763317 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2583ff3-e358-446c-9660-db7074583d8a-operator-scripts\") pod \"cloudkitty-3e4b-account-create-update-ghzkt\" (UID: \"a2583ff3-e358-446c-9660-db7074583d8a\") " pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.763398 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsccl\" (UniqueName: \"kubernetes.io/projected/a2583ff3-e358-446c-9660-db7074583d8a-kube-api-access-jsccl\") pod \"cloudkitty-3e4b-account-create-update-ghzkt\" (UID: \"a2583ff3-e358-446c-9660-db7074583d8a\") " pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.764637 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6507d8ce-f566-46d6-b55d-abbaccc71857-operator-scripts\") pod \"neutron-db-create-xlsrb\" (UID: \"6507d8ce-f566-46d6-b55d-abbaccc71857\") " pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.765392 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2583ff3-e358-446c-9660-db7074583d8a-operator-scripts\") pod \"cloudkitty-3e4b-account-create-update-ghzkt\" (UID: \"a2583ff3-e358-446c-9660-db7074583d8a\") " pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.785128 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsccl\" (UniqueName: \"kubernetes.io/projected/a2583ff3-e358-446c-9660-db7074583d8a-kube-api-access-jsccl\") pod \"cloudkitty-3e4b-account-create-update-ghzkt\" (UID: \"a2583ff3-e358-446c-9660-db7074583d8a\") " pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.786002 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t2n6\" (UniqueName: \"kubernetes.io/projected/6507d8ce-f566-46d6-b55d-abbaccc71857-kube-api-access-6t2n6\") pod \"neutron-db-create-xlsrb\" (UID: \"6507d8ce-f566-46d6-b55d-abbaccc71857\") " pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.886345 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.900660 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.916985 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:16 crc kubenswrapper[4706]: I1208 19:41:16.927844 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:17 crc kubenswrapper[4706]: I1208 19:41:17.030878 4706 generic.go:334] "Generic (PLEG): container finished" podID="ce25e3ff-7e7a-4776-a1fb-bec77f08fa56" containerID="03710859b1f1f8aede28d1c0570fe1e5ee1b7eb978dd4f839d9e220d609766bc" exitCode=0 Dec 08 19:41:17 crc kubenswrapper[4706]: I1208 19:41:17.031057 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c9ab-account-create-update-b72xz" event={"ID":"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56","Type":"ContainerDied","Data":"03710859b1f1f8aede28d1c0570fe1e5ee1b7eb978dd4f839d9e220d609766bc"} Dec 08 19:41:17 crc kubenswrapper[4706]: I1208 19:41:17.040692 4706 generic.go:334] "Generic (PLEG): container finished" podID="a36b48c2-0389-416b-a142-b28dc1ea76da" containerID="286bb523070edb6a5131de79c1e7d90a4a41362b076e9aa5d5b6b2a0e851e984" exitCode=0 Dec 08 19:41:17 crc kubenswrapper[4706]: I1208 19:41:17.040767 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pvxvl" event={"ID":"a36b48c2-0389-416b-a142-b28dc1ea76da","Type":"ContainerDied","Data":"286bb523070edb6a5131de79c1e7d90a4a41362b076e9aa5d5b6b2a0e851e984"} Dec 08 19:41:17 crc kubenswrapper[4706]: I1208 19:41:17.044804 4706 generic.go:334] "Generic (PLEG): container finished" podID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" containerID="ae486bf81a5b12911f8587408d3868244a134566a5e315c09dac14d9c7bff8e9" exitCode=0 Dec 08 19:41:17 crc kubenswrapper[4706]: I1208 19:41:17.045048 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" event={"ID":"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568","Type":"ContainerDied","Data":"ae486bf81a5b12911f8587408d3868244a134566a5e315c09dac14d9c7bff8e9"} Dec 08 19:41:17 crc kubenswrapper[4706]: I1208 19:41:17.227019 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec9a-account-create-update-bfpsl"] Dec 08 19:41:17 crc kubenswrapper[4706]: I1208 19:41:17.895706 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:17 crc kubenswrapper[4706]: E1208 19:41:17.896120 4706 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 19:41:17 crc kubenswrapper[4706]: E1208 19:41:17.896147 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 19:41:17 crc kubenswrapper[4706]: E1208 19:41:17.896212 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift podName:d81adfd7-689a-496a-b4bf-c35450b63ca9 nodeName:}" failed. No retries permitted until 2025-12-08 19:41:33.896190444 +0000 UTC m=+1176.538391447 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift") pod "swift-storage-0" (UID: "d81adfd7-689a-496a-b4bf-c35450b63ca9") : configmap "swift-ring-files" not found Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.171652 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-248t7"] Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.173243 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-248t7" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.191179 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-248t7"] Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.279925 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2e9f-account-create-update-9kkfk"] Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.287963 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.292887 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.310622 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pnxb\" (UniqueName: \"kubernetes.io/projected/3f81f35c-cf94-4180-ad69-3d54f8c41a35-kube-api-access-2pnxb\") pod \"keystone-db-create-248t7\" (UID: \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\") " pod="openstack/keystone-db-create-248t7" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.310908 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f81f35c-cf94-4180-ad69-3d54f8c41a35-operator-scripts\") pod \"keystone-db-create-248t7\" (UID: \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\") " pod="openstack/keystone-db-create-248t7" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.344317 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2e9f-account-create-update-9kkfk"] Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.404029 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-cpmsk"] Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.405771 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.416037 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pnxb\" (UniqueName: \"kubernetes.io/projected/3f81f35c-cf94-4180-ad69-3d54f8c41a35-kube-api-access-2pnxb\") pod \"keystone-db-create-248t7\" (UID: \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\") " pod="openstack/keystone-db-create-248t7" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.417417 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w8m9\" (UniqueName: \"kubernetes.io/projected/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-kube-api-access-7w8m9\") pod \"keystone-2e9f-account-create-update-9kkfk\" (UID: \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\") " pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.417521 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f81f35c-cf94-4180-ad69-3d54f8c41a35-operator-scripts\") pod \"keystone-db-create-248t7\" (UID: \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\") " pod="openstack/keystone-db-create-248t7" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.417574 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-operator-scripts\") pod \"keystone-2e9f-account-create-update-9kkfk\" (UID: \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\") " pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.418880 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f81f35c-cf94-4180-ad69-3d54f8c41a35-operator-scripts\") pod \"keystone-db-create-248t7\" (UID: \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\") " pod="openstack/keystone-db-create-248t7" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.433737 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cpmsk"] Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.493495 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pnxb\" (UniqueName: \"kubernetes.io/projected/3f81f35c-cf94-4180-ad69-3d54f8c41a35-kube-api-access-2pnxb\") pod \"keystone-db-create-248t7\" (UID: \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\") " pod="openstack/keystone-db-create-248t7" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.510831 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-248t7" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.519935 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtclq\" (UniqueName: \"kubernetes.io/projected/e1bc7a6a-8c78-4597-ba47-23e0270264e4-kube-api-access-qtclq\") pod \"placement-db-create-cpmsk\" (UID: \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\") " pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.554584 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc7a6a-8c78-4597-ba47-23e0270264e4-operator-scripts\") pod \"placement-db-create-cpmsk\" (UID: \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\") " pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.554872 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w8m9\" (UniqueName: \"kubernetes.io/projected/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-kube-api-access-7w8m9\") pod \"keystone-2e9f-account-create-update-9kkfk\" (UID: \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\") " pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.555028 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-operator-scripts\") pod \"keystone-2e9f-account-create-update-9kkfk\" (UID: \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\") " pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.555994 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-operator-scripts\") pod \"keystone-2e9f-account-create-update-9kkfk\" (UID: \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\") " pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.562095 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-04cd-account-create-update-zbfff"] Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.563963 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.566910 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.576137 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-04cd-account-create-update-zbfff"] Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.578444 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w8m9\" (UniqueName: \"kubernetes.io/projected/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-kube-api-access-7w8m9\") pod \"keystone-2e9f-account-create-update-9kkfk\" (UID: \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\") " pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.644294 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.657879 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql5vb\" (UniqueName: \"kubernetes.io/projected/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-kube-api-access-ql5vb\") pod \"placement-04cd-account-create-update-zbfff\" (UID: \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\") " pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.657993 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc7a6a-8c78-4597-ba47-23e0270264e4-operator-scripts\") pod \"placement-db-create-cpmsk\" (UID: \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\") " pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.658092 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-operator-scripts\") pod \"placement-04cd-account-create-update-zbfff\" (UID: \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\") " pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.658174 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtclq\" (UniqueName: \"kubernetes.io/projected/e1bc7a6a-8c78-4597-ba47-23e0270264e4-kube-api-access-qtclq\") pod \"placement-db-create-cpmsk\" (UID: \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\") " pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.658807 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc7a6a-8c78-4597-ba47-23e0270264e4-operator-scripts\") pod \"placement-db-create-cpmsk\" (UID: \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\") " pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.681237 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtclq\" (UniqueName: \"kubernetes.io/projected/e1bc7a6a-8c78-4597-ba47-23e0270264e4-kube-api-access-qtclq\") pod \"placement-db-create-cpmsk\" (UID: \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\") " pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.737109 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.760593 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-operator-scripts\") pod \"placement-04cd-account-create-update-zbfff\" (UID: \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\") " pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.760785 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql5vb\" (UniqueName: \"kubernetes.io/projected/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-kube-api-access-ql5vb\") pod \"placement-04cd-account-create-update-zbfff\" (UID: \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\") " pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.762711 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-operator-scripts\") pod \"placement-04cd-account-create-update-zbfff\" (UID: \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\") " pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.788038 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql5vb\" (UniqueName: \"kubernetes.io/projected/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-kube-api-access-ql5vb\") pod \"placement-04cd-account-create-update-zbfff\" (UID: \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\") " pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:18 crc kubenswrapper[4706]: I1208 19:41:18.978665 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:19 crc kubenswrapper[4706]: W1208 19:41:19.276509 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf23ba90a_3542_4241_9e27_63a9bcd2baee.slice/crio-193756302ffa7d484ff183afc8e6f30ba017ef73714ee70397951a57d95852bc WatchSource:0}: Error finding container 193756302ffa7d484ff183afc8e6f30ba017ef73714ee70397951a57d95852bc: Status 404 returned error can't find the container with id 193756302ffa7d484ff183afc8e6f30ba017ef73714ee70397951a57d95852bc Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.335086 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.340519 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.389971 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.482870 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a36b48c2-0389-416b-a142-b28dc1ea76da-operator-scripts\") pod \"a36b48c2-0389-416b-a142-b28dc1ea76da\" (UID: \"a36b48c2-0389-416b-a142-b28dc1ea76da\") " Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.483000 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-config\") pod \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.483603 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghc2w\" (UniqueName: \"kubernetes.io/projected/a36b48c2-0389-416b-a142-b28dc1ea76da-kube-api-access-ghc2w\") pod \"a36b48c2-0389-416b-a142-b28dc1ea76da\" (UID: \"a36b48c2-0389-416b-a142-b28dc1ea76da\") " Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.483708 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-dns-svc\") pod \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.483893 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz9j6\" (UniqueName: \"kubernetes.io/projected/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-kube-api-access-qz9j6\") pod \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\" (UID: \"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568\") " Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.485533 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36b48c2-0389-416b-a142-b28dc1ea76da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a36b48c2-0389-416b-a142-b28dc1ea76da" (UID: "a36b48c2-0389-416b-a142-b28dc1ea76da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.494163 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36b48c2-0389-416b-a142-b28dc1ea76da-kube-api-access-ghc2w" (OuterVolumeSpecName: "kube-api-access-ghc2w") pod "a36b48c2-0389-416b-a142-b28dc1ea76da" (UID: "a36b48c2-0389-416b-a142-b28dc1ea76da"). InnerVolumeSpecName "kube-api-access-ghc2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.499618 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-kube-api-access-qz9j6" (OuterVolumeSpecName: "kube-api-access-qz9j6") pod "52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" (UID: "52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568"). InnerVolumeSpecName "kube-api-access-qz9j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.556825 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-config" (OuterVolumeSpecName: "config") pod "52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" (UID: "52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.582952 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" (UID: "52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.586218 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8grc\" (UniqueName: \"kubernetes.io/projected/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-kube-api-access-s8grc\") pod \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\" (UID: \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\") " Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.586355 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-operator-scripts\") pod \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\" (UID: \"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56\") " Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.587199 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.587213 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz9j6\" (UniqueName: \"kubernetes.io/projected/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-kube-api-access-qz9j6\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.587228 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a36b48c2-0389-416b-a142-b28dc1ea76da-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.587240 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.587249 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghc2w\" (UniqueName: \"kubernetes.io/projected/a36b48c2-0389-416b-a142-b28dc1ea76da-kube-api-access-ghc2w\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.587582 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce25e3ff-7e7a-4776-a1fb-bec77f08fa56" (UID: "ce25e3ff-7e7a-4776-a1fb-bec77f08fa56"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.621549 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-kube-api-access-s8grc" (OuterVolumeSpecName: "kube-api-access-s8grc") pod "ce25e3ff-7e7a-4776-a1fb-bec77f08fa56" (UID: "ce25e3ff-7e7a-4776-a1fb-bec77f08fa56"). InnerVolumeSpecName "kube-api-access-s8grc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.691055 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8grc\" (UniqueName: \"kubernetes.io/projected/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-kube-api-access-s8grc\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.691099 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.788590 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:41:19 crc kubenswrapper[4706]: I1208 19:41:19.810752 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fb6gx" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.122296 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-78wps-config-hl42b"] Dec 08 19:41:20 crc kubenswrapper[4706]: E1208 19:41:20.123337 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" containerName="dnsmasq-dns" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.123359 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" containerName="dnsmasq-dns" Dec 08 19:41:20 crc kubenswrapper[4706]: E1208 19:41:20.123381 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce25e3ff-7e7a-4776-a1fb-bec77f08fa56" containerName="mariadb-account-create-update" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.123391 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce25e3ff-7e7a-4776-a1fb-bec77f08fa56" containerName="mariadb-account-create-update" Dec 08 19:41:20 crc kubenswrapper[4706]: E1208 19:41:20.123411 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" containerName="init" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.123419 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" containerName="init" Dec 08 19:41:20 crc kubenswrapper[4706]: E1208 19:41:20.123453 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36b48c2-0389-416b-a142-b28dc1ea76da" containerName="mariadb-database-create" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.123462 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36b48c2-0389-416b-a142-b28dc1ea76da" containerName="mariadb-database-create" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.123716 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" containerName="dnsmasq-dns" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.123737 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce25e3ff-7e7a-4776-a1fb-bec77f08fa56" containerName="mariadb-account-create-update" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.123762 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36b48c2-0389-416b-a142-b28dc1ea76da" containerName="mariadb-database-create" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.124808 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.134775 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.150795 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78wps-config-hl42b"] Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.162611 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pvxvl" event={"ID":"a36b48c2-0389-416b-a142-b28dc1ea76da","Type":"ContainerDied","Data":"6f9e6eff7aa050660d3c8b7cf8cf1732d2a1b09d9bccf03e02b4360b88b7a2c7"} Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.162655 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f9e6eff7aa050660d3c8b7cf8cf1732d2a1b09d9bccf03e02b4360b88b7a2c7" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.162729 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pvxvl" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.176212 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.177152 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lqbnv" event={"ID":"52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568","Type":"ContainerDied","Data":"45e0293d8502ead0e496e9839adaca9aaaea39509c8f6860e7935e13cc20f8f4"} Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.177386 4706 scope.go:117] "RemoveContainer" containerID="ae486bf81a5b12911f8587408d3868244a134566a5e315c09dac14d9c7bff8e9" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.186586 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec9a-account-create-update-bfpsl" event={"ID":"f23ba90a-3542-4241-9e27-63a9bcd2baee","Type":"ContainerStarted","Data":"f14ffbe72b87da1bf06aed478ace4a9bd69b5cd72f7321aa5d834315bfcbdb29"} Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.186630 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec9a-account-create-update-bfpsl" event={"ID":"f23ba90a-3542-4241-9e27-63a9bcd2baee","Type":"ContainerStarted","Data":"193756302ffa7d484ff183afc8e6f30ba017ef73714ee70397951a57d95852bc"} Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.199347 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerStarted","Data":"22c94ccf73de5e86f0282d36e79d17564319da67bf2ad260d8fdbb3dee34a26d"} Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.211851 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c9ab-account-create-update-b72xz" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.211961 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c9ab-account-create-update-b72xz" event={"ID":"ce25e3ff-7e7a-4776-a1fb-bec77f08fa56","Type":"ContainerDied","Data":"8b2daf4c0fa100aa65f11cd97e7fd79a4b53132309f2635206cdfa8e89dfb1a2"} Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.211994 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b2daf4c0fa100aa65f11cd97e7fd79a4b53132309f2635206cdfa8e89dfb1a2" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.214130 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec9a-account-create-update-bfpsl" podStartSLOduration=5.214099128 podStartE2EDuration="5.214099128s" podCreationTimestamp="2025-12-08 19:41:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:20.206131271 +0000 UTC m=+1162.848332274" watchObservedRunningTime="2025-12-08 19:41:20.214099128 +0000 UTC m=+1162.856300131" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.216921 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-log-ovn\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.217078 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-additional-scripts\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.217133 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-scripts\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.217401 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.217543 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run-ovn\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.217638 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sssd\" (UniqueName: \"kubernetes.io/projected/d1ae1b42-40e1-4593-9068-4f0319897bd4-kube-api-access-5sssd\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.252462 4706 scope.go:117] "RemoveContainer" containerID="f4312efa40125432255026a9fc1254f221ccfdb2a54152a5d299b4351291faa2" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.294725 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=25.240930135 podStartE2EDuration="1m20.294697441s" podCreationTimestamp="2025-12-08 19:40:00 +0000 UTC" firstStartedPulling="2025-12-08 19:40:24.512255858 +0000 UTC m=+1107.154456861" lastFinishedPulling="2025-12-08 19:41:19.566023164 +0000 UTC m=+1162.208224167" observedRunningTime="2025-12-08 19:41:20.24898415 +0000 UTC m=+1162.891185163" watchObservedRunningTime="2025-12-08 19:41:20.294697441 +0000 UTC m=+1162.936898444" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.299407 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lqbnv"] Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.310840 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lqbnv"] Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.321755 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.321979 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run-ovn\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.322065 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sssd\" (UniqueName: \"kubernetes.io/projected/d1ae1b42-40e1-4593-9068-4f0319897bd4-kube-api-access-5sssd\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.322122 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-log-ovn\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.322190 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.323141 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-additional-scripts\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.322204 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-additional-scripts\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.323197 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-log-ovn\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.323289 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-scripts\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.324198 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run-ovn\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.327646 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-scripts\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.343110 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-szsrk"] Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.348720 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sssd\" (UniqueName: \"kubernetes.io/projected/d1ae1b42-40e1-4593-9068-4f0319897bd4-kube-api-access-5sssd\") pod \"ovn-controller-78wps-config-hl42b\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: I1208 19:41:20.353881 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:20 crc kubenswrapper[4706]: W1208 19:41:20.406068 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod091f4a40_69b6_4c09_8251_d5aca75a420f.slice/crio-1a6a627738232c915ca13d2db771d8d10701308e8fe0e7ff7c3f7ef1a7745475 WatchSource:0}: Error finding container 1a6a627738232c915ca13d2db771d8d10701308e8fe0e7ff7c3f7ef1a7745475: Status 404 returned error can't find the container with id 1a6a627738232c915ca13d2db771d8d10701308e8fe0e7ff7c3f7ef1a7745475 Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.231649 4706 generic.go:334] "Generic (PLEG): container finished" podID="f23ba90a-3542-4241-9e27-63a9bcd2baee" containerID="f14ffbe72b87da1bf06aed478ace4a9bd69b5cd72f7321aa5d834315bfcbdb29" exitCode=0 Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.232202 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec9a-account-create-update-bfpsl" event={"ID":"f23ba90a-3542-4241-9e27-63a9bcd2baee","Type":"ContainerDied","Data":"f14ffbe72b87da1bf06aed478ace4a9bd69b5cd72f7321aa5d834315bfcbdb29"} Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.235233 4706 generic.go:334] "Generic (PLEG): container finished" podID="091f4a40-69b6-4c09-8251-d5aca75a420f" containerID="9a3436d1e435ccfa19264d07fdaa4e33e7ef3a978b4748d07a75554255977370" exitCode=0 Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.236547 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-szsrk" event={"ID":"091f4a40-69b6-4c09-8251-d5aca75a420f","Type":"ContainerDied","Data":"9a3436d1e435ccfa19264d07fdaa4e33e7ef3a978b4748d07a75554255977370"} Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.236573 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-szsrk" event={"ID":"091f4a40-69b6-4c09-8251-d5aca75a420f","Type":"ContainerStarted","Data":"1a6a627738232c915ca13d2db771d8d10701308e8fe0e7ff7c3f7ef1a7745475"} Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.291319 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-248t7"] Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.303749 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-6lf8v"] Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.313577 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xlsrb"] Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.325172 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dfae-account-create-update-dhrlq"] Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.374190 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-pjffv"] Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.399544 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b7de-account-create-update-ntz9l"] Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.415497 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2e9f-account-create-update-9kkfk"] Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.433797 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-3e4b-account-create-update-ghzkt"] Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.441546 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cpmsk"] Dec 08 19:41:21 crc kubenswrapper[4706]: W1208 19:41:21.450293 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a7451c2_8247_406f_9d8f_eb5690a16555.slice/crio-8d5042e8a09a720b789315b7bd661ad108c4fee5344eb081e02d9eaa24611707 WatchSource:0}: Error finding container 8d5042e8a09a720b789315b7bd661ad108c4fee5344eb081e02d9eaa24611707: Status 404 returned error can't find the container with id 8d5042e8a09a720b789315b7bd661ad108c4fee5344eb081e02d9eaa24611707 Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.453535 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-04cd-account-create-update-zbfff"] Dec 08 19:41:21 crc kubenswrapper[4706]: W1208 19:41:21.464831 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb919dfd_81a4_42b1_8e30_4788113753d2.slice/crio-5bf0adf855b4bd3600f2731765ef469fb3738c5bcdb02d3dc92bd888576fea3a WatchSource:0}: Error finding container 5bf0adf855b4bd3600f2731765ef469fb3738c5bcdb02d3dc92bd888576fea3a: Status 404 returned error can't find the container with id 5bf0adf855b4bd3600f2731765ef469fb3738c5bcdb02d3dc92bd888576fea3a Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.465635 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78wps-config-hl42b"] Dec 08 19:41:21 crc kubenswrapper[4706]: W1208 19:41:21.505859 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1bc7a6a_8c78_4597_ba47_23e0270264e4.slice/crio-0a31e25e5f16d73523ace2ff9a1308b2a6407434f80b8db73ceae720010c8c47 WatchSource:0}: Error finding container 0a31e25e5f16d73523ace2ff9a1308b2a6407434f80b8db73ceae720010c8c47: Status 404 returned error can't find the container with id 0a31e25e5f16d73523ace2ff9a1308b2a6407434f80b8db73ceae720010c8c47 Dec 08 19:41:21 crc kubenswrapper[4706]: W1208 19:41:21.531585 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1ae1b42_40e1_4593_9068_4f0319897bd4.slice/crio-f6ec4cde8d20a962d96c45ec71746d8a219353e2833f5ee46bed4ae871859b12 WatchSource:0}: Error finding container f6ec4cde8d20a962d96c45ec71746d8a219353e2833f5ee46bed4ae871859b12: Status 404 returned error can't find the container with id f6ec4cde8d20a962d96c45ec71746d8a219353e2833f5ee46bed4ae871859b12 Dec 08 19:41:21 crc kubenswrapper[4706]: I1208 19:41:21.624825 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568" path="/var/lib/kubelet/pods/52a95e2d-d2d3-4a2e-9e4a-bd3238ed9568/volumes" Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.060032 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.250163 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-04cd-account-create-update-zbfff" event={"ID":"6d1a2245-a7e5-4f5f-a392-da5c9aec368a","Type":"ContainerStarted","Data":"36b7a3cf54674937b1fdb1075a42076156d95f4f8f42b722857eb3de86488575"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.252453 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cpmsk" event={"ID":"e1bc7a6a-8c78-4597-ba47-23e0270264e4","Type":"ContainerStarted","Data":"43e0453592947499e12f422ebc2fc68b6ddf6320758593f1867e1a82293cc3c6"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.252487 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cpmsk" event={"ID":"e1bc7a6a-8c78-4597-ba47-23e0270264e4","Type":"ContainerStarted","Data":"0a31e25e5f16d73523ace2ff9a1308b2a6407434f80b8db73ceae720010c8c47"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.255857 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78wps-config-hl42b" event={"ID":"d1ae1b42-40e1-4593-9068-4f0319897bd4","Type":"ContainerStarted","Data":"f6ec4cde8d20a962d96c45ec71746d8a219353e2833f5ee46bed4ae871859b12"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.258879 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dfae-account-create-update-dhrlq" event={"ID":"53d89be3-c6e5-4f3e-b788-4811b47ba332","Type":"ContainerStarted","Data":"481f9f540680ba35cb6cbf21b26f617bc25d9d5fce4bba116507a1ea39fcd8d4"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.258923 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dfae-account-create-update-dhrlq" event={"ID":"53d89be3-c6e5-4f3e-b788-4811b47ba332","Type":"ContainerStarted","Data":"34c9234c22ac0a523ff5403fe57780eb967337c285703437fb4a35d898e017e1"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.263087 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xlsrb" event={"ID":"6507d8ce-f566-46d6-b55d-abbaccc71857","Type":"ContainerStarted","Data":"f8fb649ccf58b41a0af70194c43aff3468d21ddcad19af964b7e4496056ed6dc"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.263136 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xlsrb" event={"ID":"6507d8ce-f566-46d6-b55d-abbaccc71857","Type":"ContainerStarted","Data":"277a7137bce66705b07fdf26feed278d8780b499d0b8ef8c6c60c1d43a8508c9"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.266404 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-248t7" event={"ID":"3f81f35c-cf94-4180-ad69-3d54f8c41a35","Type":"ContainerStarted","Data":"2796983a054cd5070dabaa79457a6d5db52c03892764167985d2edb3a6242898"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.266435 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-248t7" event={"ID":"3f81f35c-cf94-4180-ad69-3d54f8c41a35","Type":"ContainerStarted","Data":"bcd44ec47757503b1f466666c79295b22283a2a3d65f6ca391d57c64235f1eb8"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.269537 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" event={"ID":"a2583ff3-e358-446c-9660-db7074583d8a","Type":"ContainerStarted","Data":"5cd2f220ebfc90e073ad3511d3758d1ceb9149e44e8b3974bf5b5ec04cbf4d91"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.272546 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-cpmsk" podStartSLOduration=4.272531898 podStartE2EDuration="4.272531898s" podCreationTimestamp="2025-12-08 19:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:22.270048888 +0000 UTC m=+1164.912249891" watchObservedRunningTime="2025-12-08 19:41:22.272531898 +0000 UTC m=+1164.914732901" Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.274689 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2e9f-account-create-update-9kkfk" event={"ID":"fd2c1239-cb7e-4bc1-af35-75fcd1613c77","Type":"ContainerStarted","Data":"600d5def7d38461ce26b35d9038face2cc040047cebb58b0ade2fefcaad591ef"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.274753 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2e9f-account-create-update-9kkfk" event={"ID":"fd2c1239-cb7e-4bc1-af35-75fcd1613c77","Type":"ContainerStarted","Data":"1c1123a563c8838fb69c2ce676979791184c7836ee37129759387fe5d1a2d47c"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.278871 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pjffv" event={"ID":"bb919dfd-81a4-42b1-8e30-4788113753d2","Type":"ContainerStarted","Data":"89fd8dc0f0a6ef5e3f3f07e8d9cdefed1911f9274b9723c5165549b21a5a9e77"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.278904 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pjffv" event={"ID":"bb919dfd-81a4-42b1-8e30-4788113753d2","Type":"ContainerStarted","Data":"5bf0adf855b4bd3600f2731765ef469fb3738c5bcdb02d3dc92bd888576fea3a"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.335044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b7de-account-create-update-ntz9l" event={"ID":"5a7451c2-8247-406f-9d8f-eb5690a16555","Type":"ContainerStarted","Data":"af4da9490fac94bb822be0a586032bbf3570469c690e73984ee6d116c975e496"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.335134 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b7de-account-create-update-ntz9l" event={"ID":"5a7451c2-8247-406f-9d8f-eb5690a16555","Type":"ContainerStarted","Data":"8d5042e8a09a720b789315b7bd661ad108c4fee5344eb081e02d9eaa24611707"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.362814 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dfae-account-create-update-dhrlq" podStartSLOduration=6.362785267 podStartE2EDuration="6.362785267s" podCreationTimestamp="2025-12-08 19:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:22.309752177 +0000 UTC m=+1164.951953200" watchObservedRunningTime="2025-12-08 19:41:22.362785267 +0000 UTC m=+1165.004986270" Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.372839 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-6lf8v" event={"ID":"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a","Type":"ContainerStarted","Data":"410357fb3263e326acd30036fd69bcf7fe0c2b8a434de30b3506236433235dba"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.372892 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-6lf8v" event={"ID":"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a","Type":"ContainerStarted","Data":"706f5fb94216b1b0f593b8b41d8f3c7f65b88628528f2e2fb7264533690f638f"} Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.417090 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-248t7" podStartSLOduration=4.417063191 podStartE2EDuration="4.417063191s" podCreationTimestamp="2025-12-08 19:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:22.326835194 +0000 UTC m=+1164.969036197" watchObservedRunningTime="2025-12-08 19:41:22.417063191 +0000 UTC m=+1165.059264194" Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.448929 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-xlsrb" podStartSLOduration=6.448897217 podStartE2EDuration="6.448897217s" podCreationTimestamp="2025-12-08 19:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:22.359628827 +0000 UTC m=+1165.001829830" watchObservedRunningTime="2025-12-08 19:41:22.448897217 +0000 UTC m=+1165.091098220" Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.460201 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-2e9f-account-create-update-9kkfk" podStartSLOduration=4.460013074 podStartE2EDuration="4.460013074s" podCreationTimestamp="2025-12-08 19:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:22.387950513 +0000 UTC m=+1165.030151516" watchObservedRunningTime="2025-12-08 19:41:22.460013074 +0000 UTC m=+1165.102214077" Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.484777 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-b7de-account-create-update-ntz9l" podStartSLOduration=6.484744248 podStartE2EDuration="6.484744248s" podCreationTimestamp="2025-12-08 19:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:22.413714826 +0000 UTC m=+1165.055915829" watchObservedRunningTime="2025-12-08 19:41:22.484744248 +0000 UTC m=+1165.126945251" Dec 08 19:41:22 crc kubenswrapper[4706]: I1208 19:41:22.492100 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-create-6lf8v" podStartSLOduration=6.492078846 podStartE2EDuration="6.492078846s" podCreationTimestamp="2025-12-08 19:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:22.445748308 +0000 UTC m=+1165.087949311" watchObservedRunningTime="2025-12-08 19:41:22.492078846 +0000 UTC m=+1165.134279849" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.256782 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.266227 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.354203 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmwzd\" (UniqueName: \"kubernetes.io/projected/091f4a40-69b6-4c09-8251-d5aca75a420f-kube-api-access-rmwzd\") pod \"091f4a40-69b6-4c09-8251-d5aca75a420f\" (UID: \"091f4a40-69b6-4c09-8251-d5aca75a420f\") " Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.354365 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f23ba90a-3542-4241-9e27-63a9bcd2baee-operator-scripts\") pod \"f23ba90a-3542-4241-9e27-63a9bcd2baee\" (UID: \"f23ba90a-3542-4241-9e27-63a9bcd2baee\") " Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.354509 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/091f4a40-69b6-4c09-8251-d5aca75a420f-operator-scripts\") pod \"091f4a40-69b6-4c09-8251-d5aca75a420f\" (UID: \"091f4a40-69b6-4c09-8251-d5aca75a420f\") " Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.354546 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r28zg\" (UniqueName: \"kubernetes.io/projected/f23ba90a-3542-4241-9e27-63a9bcd2baee-kube-api-access-r28zg\") pod \"f23ba90a-3542-4241-9e27-63a9bcd2baee\" (UID: \"f23ba90a-3542-4241-9e27-63a9bcd2baee\") " Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.355458 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091f4a40-69b6-4c09-8251-d5aca75a420f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "091f4a40-69b6-4c09-8251-d5aca75a420f" (UID: "091f4a40-69b6-4c09-8251-d5aca75a420f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.355480 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f23ba90a-3542-4241-9e27-63a9bcd2baee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f23ba90a-3542-4241-9e27-63a9bcd2baee" (UID: "f23ba90a-3542-4241-9e27-63a9bcd2baee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.372388 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091f4a40-69b6-4c09-8251-d5aca75a420f-kube-api-access-rmwzd" (OuterVolumeSpecName: "kube-api-access-rmwzd") pod "091f4a40-69b6-4c09-8251-d5aca75a420f" (UID: "091f4a40-69b6-4c09-8251-d5aca75a420f"). InnerVolumeSpecName "kube-api-access-rmwzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.372620 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f23ba90a-3542-4241-9e27-63a9bcd2baee-kube-api-access-r28zg" (OuterVolumeSpecName: "kube-api-access-r28zg") pod "f23ba90a-3542-4241-9e27-63a9bcd2baee" (UID: "f23ba90a-3542-4241-9e27-63a9bcd2baee"). InnerVolumeSpecName "kube-api-access-r28zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.392524 4706 generic.go:334] "Generic (PLEG): container finished" podID="3f81f35c-cf94-4180-ad69-3d54f8c41a35" containerID="2796983a054cd5070dabaa79457a6d5db52c03892764167985d2edb3a6242898" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.392601 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-248t7" event={"ID":"3f81f35c-cf94-4180-ad69-3d54f8c41a35","Type":"ContainerDied","Data":"2796983a054cd5070dabaa79457a6d5db52c03892764167985d2edb3a6242898"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.400069 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1ae1b42-40e1-4593-9068-4f0319897bd4" containerID="927f5c41ed95fe2418cd7edb1c7e27514c71e951fd9dc76c0b08c1bd76c67e93" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.400161 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78wps-config-hl42b" event={"ID":"d1ae1b42-40e1-4593-9068-4f0319897bd4","Type":"ContainerDied","Data":"927f5c41ed95fe2418cd7edb1c7e27514c71e951fd9dc76c0b08c1bd76c67e93"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.405232 4706 generic.go:334] "Generic (PLEG): container finished" podID="a2583ff3-e358-446c-9660-db7074583d8a" containerID="1e35ee43159ba60bac417a2197bb3ac0f6c9d3ca476a87ce1526580a19b706da" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.405322 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" event={"ID":"a2583ff3-e358-446c-9660-db7074583d8a","Type":"ContainerDied","Data":"1e35ee43159ba60bac417a2197bb3ac0f6c9d3ca476a87ce1526580a19b706da"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.408364 4706 generic.go:334] "Generic (PLEG): container finished" podID="6d1a2245-a7e5-4f5f-a392-da5c9aec368a" containerID="17c235c39e4638c7623d7a98b4579a281bb377c5ce7a151855daa8e97f18c37e" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.408412 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-04cd-account-create-update-zbfff" event={"ID":"6d1a2245-a7e5-4f5f-a392-da5c9aec368a","Type":"ContainerDied","Data":"17c235c39e4638c7623d7a98b4579a281bb377c5ce7a151855daa8e97f18c37e"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.410253 4706 generic.go:334] "Generic (PLEG): container finished" podID="fd2c1239-cb7e-4bc1-af35-75fcd1613c77" containerID="600d5def7d38461ce26b35d9038face2cc040047cebb58b0ade2fefcaad591ef" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.410331 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2e9f-account-create-update-9kkfk" event={"ID":"fd2c1239-cb7e-4bc1-af35-75fcd1613c77","Type":"ContainerDied","Data":"600d5def7d38461ce26b35d9038face2cc040047cebb58b0ade2fefcaad591ef"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.415886 4706 generic.go:334] "Generic (PLEG): container finished" podID="5a7451c2-8247-406f-9d8f-eb5690a16555" containerID="af4da9490fac94bb822be0a586032bbf3570469c690e73984ee6d116c975e496" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.416134 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b7de-account-create-update-ntz9l" event={"ID":"5a7451c2-8247-406f-9d8f-eb5690a16555","Type":"ContainerDied","Data":"af4da9490fac94bb822be0a586032bbf3570469c690e73984ee6d116c975e496"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.418852 4706 generic.go:334] "Generic (PLEG): container finished" podID="53d89be3-c6e5-4f3e-b788-4811b47ba332" containerID="481f9f540680ba35cb6cbf21b26f617bc25d9d5fce4bba116507a1ea39fcd8d4" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.418898 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dfae-account-create-update-dhrlq" event={"ID":"53d89be3-c6e5-4f3e-b788-4811b47ba332","Type":"ContainerDied","Data":"481f9f540680ba35cb6cbf21b26f617bc25d9d5fce4bba116507a1ea39fcd8d4"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.424348 4706 generic.go:334] "Generic (PLEG): container finished" podID="6507d8ce-f566-46d6-b55d-abbaccc71857" containerID="f8fb649ccf58b41a0af70194c43aff3468d21ddcad19af964b7e4496056ed6dc" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.424429 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xlsrb" event={"ID":"6507d8ce-f566-46d6-b55d-abbaccc71857","Type":"ContainerDied","Data":"f8fb649ccf58b41a0af70194c43aff3468d21ddcad19af964b7e4496056ed6dc"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.426565 4706 generic.go:334] "Generic (PLEG): container finished" podID="bb919dfd-81a4-42b1-8e30-4788113753d2" containerID="89fd8dc0f0a6ef5e3f3f07e8d9cdefed1911f9274b9723c5165549b21a5a9e77" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.426615 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pjffv" event={"ID":"bb919dfd-81a4-42b1-8e30-4788113753d2","Type":"ContainerDied","Data":"89fd8dc0f0a6ef5e3f3f07e8d9cdefed1911f9274b9723c5165549b21a5a9e77"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.444424 4706 generic.go:334] "Generic (PLEG): container finished" podID="e1bc7a6a-8c78-4597-ba47-23e0270264e4" containerID="43e0453592947499e12f422ebc2fc68b6ddf6320758593f1867e1a82293cc3c6" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.444474 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cpmsk" event={"ID":"e1bc7a6a-8c78-4597-ba47-23e0270264e4","Type":"ContainerDied","Data":"43e0453592947499e12f422ebc2fc68b6ddf6320758593f1867e1a82293cc3c6"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.450486 4706 generic.go:334] "Generic (PLEG): container finished" podID="50a60ccd-c961-4c0c-ae74-0b0d9c6e172a" containerID="410357fb3263e326acd30036fd69bcf7fe0c2b8a434de30b3506236433235dba" exitCode=0 Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.450609 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-6lf8v" event={"ID":"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a","Type":"ContainerDied","Data":"410357fb3263e326acd30036fd69bcf7fe0c2b8a434de30b3506236433235dba"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.455806 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec9a-account-create-update-bfpsl" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.456632 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec9a-account-create-update-bfpsl" event={"ID":"f23ba90a-3542-4241-9e27-63a9bcd2baee","Type":"ContainerDied","Data":"193756302ffa7d484ff183afc8e6f30ba017ef73714ee70397951a57d95852bc"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.456703 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="193756302ffa7d484ff183afc8e6f30ba017ef73714ee70397951a57d95852bc" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.457467 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmwzd\" (UniqueName: \"kubernetes.io/projected/091f4a40-69b6-4c09-8251-d5aca75a420f-kube-api-access-rmwzd\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.457495 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f23ba90a-3542-4241-9e27-63a9bcd2baee-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.457505 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/091f4a40-69b6-4c09-8251-d5aca75a420f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.457516 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r28zg\" (UniqueName: \"kubernetes.io/projected/f23ba90a-3542-4241-9e27-63a9bcd2baee-kube-api-access-r28zg\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.459564 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-szsrk" event={"ID":"091f4a40-69b6-4c09-8251-d5aca75a420f","Type":"ContainerDied","Data":"1a6a627738232c915ca13d2db771d8d10701308e8fe0e7ff7c3f7ef1a7745475"} Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.459605 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a6a627738232c915ca13d2db771d8d10701308e8fe0e7ff7c3f7ef1a7745475" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.459627 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-szsrk" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.895346 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="9a912025-8f8e-4d61-861e-c933dafe4660" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.961598 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-8s7bd"] Dec 08 19:41:23 crc kubenswrapper[4706]: E1208 19:41:23.962108 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091f4a40-69b6-4c09-8251-d5aca75a420f" containerName="mariadb-database-create" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.962129 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="091f4a40-69b6-4c09-8251-d5aca75a420f" containerName="mariadb-database-create" Dec 08 19:41:23 crc kubenswrapper[4706]: E1208 19:41:23.962158 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f23ba90a-3542-4241-9e27-63a9bcd2baee" containerName="mariadb-account-create-update" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.962167 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f23ba90a-3542-4241-9e27-63a9bcd2baee" containerName="mariadb-account-create-update" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.962510 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="091f4a40-69b6-4c09-8251-d5aca75a420f" containerName="mariadb-database-create" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.962538 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f23ba90a-3542-4241-9e27-63a9bcd2baee" containerName="mariadb-account-create-update" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.963300 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.970500 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kvdqs" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.970720 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 08 19:41:23 crc kubenswrapper[4706]: I1208 19:41:23.972718 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8s7bd"] Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.073602 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-db-sync-config-data\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.073660 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fv45\" (UniqueName: \"kubernetes.io/projected/e18991e1-84d5-4e32-9a05-ad08f31cbd60-kube-api-access-9fv45\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.073806 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-combined-ca-bundle\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.073893 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-config-data\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.175808 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-config-data\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.175965 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-db-sync-config-data\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.176008 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fv45\" (UniqueName: \"kubernetes.io/projected/e18991e1-84d5-4e32-9a05-ad08f31cbd60-kube-api-access-9fv45\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.176159 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-combined-ca-bundle\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.182504 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-db-sync-config-data\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.181871 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-config-data\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.184384 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-combined-ca-bundle\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.202782 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fv45\" (UniqueName: \"kubernetes.io/projected/e18991e1-84d5-4e32-9a05-ad08f31cbd60-kube-api-access-9fv45\") pod \"glance-db-sync-8s7bd\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.283774 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8s7bd" Dec 08 19:41:24 crc kubenswrapper[4706]: I1208 19:41:24.557275 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-78wps" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.603846 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cpmsk" event={"ID":"e1bc7a6a-8c78-4597-ba47-23e0270264e4","Type":"ContainerDied","Data":"0a31e25e5f16d73523ace2ff9a1308b2a6407434f80b8db73ceae720010c8c47"} Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.604231 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a31e25e5f16d73523ace2ff9a1308b2a6407434f80b8db73ceae720010c8c47" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.845805 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.858975 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.875694 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.875711 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t2n6\" (UniqueName: \"kubernetes.io/projected/6507d8ce-f566-46d6-b55d-abbaccc71857-kube-api-access-6t2n6\") pod \"6507d8ce-f566-46d6-b55d-abbaccc71857\" (UID: \"6507d8ce-f566-46d6-b55d-abbaccc71857\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.875910 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc7a6a-8c78-4597-ba47-23e0270264e4-operator-scripts\") pod \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\" (UID: \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.875938 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtclq\" (UniqueName: \"kubernetes.io/projected/e1bc7a6a-8c78-4597-ba47-23e0270264e4-kube-api-access-qtclq\") pod \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\" (UID: \"e1bc7a6a-8c78-4597-ba47-23e0270264e4\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.875974 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6507d8ce-f566-46d6-b55d-abbaccc71857-operator-scripts\") pod \"6507d8ce-f566-46d6-b55d-abbaccc71857\" (UID: \"6507d8ce-f566-46d6-b55d-abbaccc71857\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.877358 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6507d8ce-f566-46d6-b55d-abbaccc71857-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6507d8ce-f566-46d6-b55d-abbaccc71857" (UID: "6507d8ce-f566-46d6-b55d-abbaccc71857"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.877867 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1bc7a6a-8c78-4597-ba47-23e0270264e4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1bc7a6a-8c78-4597-ba47-23e0270264e4" (UID: "e1bc7a6a-8c78-4597-ba47-23e0270264e4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.884937 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1bc7a6a-8c78-4597-ba47-23e0270264e4-kube-api-access-qtclq" (OuterVolumeSpecName: "kube-api-access-qtclq") pod "e1bc7a6a-8c78-4597-ba47-23e0270264e4" (UID: "e1bc7a6a-8c78-4597-ba47-23e0270264e4"). InnerVolumeSpecName "kube-api-access-qtclq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.891551 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6507d8ce-f566-46d6-b55d-abbaccc71857-kube-api-access-6t2n6" (OuterVolumeSpecName: "kube-api-access-6t2n6") pod "6507d8ce-f566-46d6-b55d-abbaccc71857" (UID: "6507d8ce-f566-46d6-b55d-abbaccc71857"). InnerVolumeSpecName "kube-api-access-6t2n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.910338 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.956670 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.964922 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.970745 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977575 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run\") pod \"d1ae1b42-40e1-4593-9068-4f0319897bd4\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977667 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-log-ovn\") pod \"d1ae1b42-40e1-4593-9068-4f0319897bd4\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977742 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53d89be3-c6e5-4f3e-b788-4811b47ba332-operator-scripts\") pod \"53d89be3-c6e5-4f3e-b788-4811b47ba332\" (UID: \"53d89be3-c6e5-4f3e-b788-4811b47ba332\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977771 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h475g\" (UniqueName: \"kubernetes.io/projected/5a7451c2-8247-406f-9d8f-eb5690a16555-kube-api-access-h475g\") pod \"5a7451c2-8247-406f-9d8f-eb5690a16555\" (UID: \"5a7451c2-8247-406f-9d8f-eb5690a16555\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977819 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-operator-scripts\") pod \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\" (UID: \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977901 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sssd\" (UniqueName: \"kubernetes.io/projected/d1ae1b42-40e1-4593-9068-4f0319897bd4-kube-api-access-5sssd\") pod \"d1ae1b42-40e1-4593-9068-4f0319897bd4\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977906 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d1ae1b42-40e1-4593-9068-4f0319897bd4" (UID: "d1ae1b42-40e1-4593-9068-4f0319897bd4"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977919 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-additional-scripts\") pod \"d1ae1b42-40e1-4593-9068-4f0319897bd4\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977967 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run-ovn\") pod \"d1ae1b42-40e1-4593-9068-4f0319897bd4\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.977996 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7451c2-8247-406f-9d8f-eb5690a16555-operator-scripts\") pod \"5a7451c2-8247-406f-9d8f-eb5690a16555\" (UID: \"5a7451c2-8247-406f-9d8f-eb5690a16555\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978040 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trbwj\" (UniqueName: \"kubernetes.io/projected/53d89be3-c6e5-4f3e-b788-4811b47ba332-kube-api-access-trbwj\") pod \"53d89be3-c6e5-4f3e-b788-4811b47ba332\" (UID: \"53d89be3-c6e5-4f3e-b788-4811b47ba332\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978078 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-scripts\") pod \"d1ae1b42-40e1-4593-9068-4f0319897bd4\" (UID: \"d1ae1b42-40e1-4593-9068-4f0319897bd4\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978102 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nlzj\" (UniqueName: \"kubernetes.io/projected/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-kube-api-access-5nlzj\") pod \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\" (UID: \"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a\") " Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978669 4706 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978682 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc7a6a-8c78-4597-ba47-23e0270264e4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978696 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtclq\" (UniqueName: \"kubernetes.io/projected/e1bc7a6a-8c78-4597-ba47-23e0270264e4-kube-api-access-qtclq\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978707 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6507d8ce-f566-46d6-b55d-abbaccc71857-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978721 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t2n6\" (UniqueName: \"kubernetes.io/projected/6507d8ce-f566-46d6-b55d-abbaccc71857-kube-api-access-6t2n6\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978933 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d1ae1b42-40e1-4593-9068-4f0319897bd4" (UID: "d1ae1b42-40e1-4593-9068-4f0319897bd4"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.978996 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run" (OuterVolumeSpecName: "var-run") pod "d1ae1b42-40e1-4593-9068-4f0319897bd4" (UID: "d1ae1b42-40e1-4593-9068-4f0319897bd4"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.979039 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d1ae1b42-40e1-4593-9068-4f0319897bd4" (UID: "d1ae1b42-40e1-4593-9068-4f0319897bd4"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.979516 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a7451c2-8247-406f-9d8f-eb5690a16555-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a7451c2-8247-406f-9d8f-eb5690a16555" (UID: "5a7451c2-8247-406f-9d8f-eb5690a16555"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.981080 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d89be3-c6e5-4f3e-b788-4811b47ba332-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53d89be3-c6e5-4f3e-b788-4811b47ba332" (UID: "53d89be3-c6e5-4f3e-b788-4811b47ba332"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.984633 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "50a60ccd-c961-4c0c-ae74-0b0d9c6e172a" (UID: "50a60ccd-c961-4c0c-ae74-0b0d9c6e172a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.986312 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.986322 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a7451c2-8247-406f-9d8f-eb5690a16555-kube-api-access-h475g" (OuterVolumeSpecName: "kube-api-access-h475g") pod "5a7451c2-8247-406f-9d8f-eb5690a16555" (UID: "5a7451c2-8247-406f-9d8f-eb5690a16555"). InnerVolumeSpecName "kube-api-access-h475g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.988682 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-scripts" (OuterVolumeSpecName: "scripts") pod "d1ae1b42-40e1-4593-9068-4f0319897bd4" (UID: "d1ae1b42-40e1-4593-9068-4f0319897bd4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.988858 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d89be3-c6e5-4f3e-b788-4811b47ba332-kube-api-access-trbwj" (OuterVolumeSpecName: "kube-api-access-trbwj") pod "53d89be3-c6e5-4f3e-b788-4811b47ba332" (UID: "53d89be3-c6e5-4f3e-b788-4811b47ba332"). InnerVolumeSpecName "kube-api-access-trbwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.991362 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ae1b42-40e1-4593-9068-4f0319897bd4-kube-api-access-5sssd" (OuterVolumeSpecName: "kube-api-access-5sssd") pod "d1ae1b42-40e1-4593-9068-4f0319897bd4" (UID: "d1ae1b42-40e1-4593-9068-4f0319897bd4"). InnerVolumeSpecName "kube-api-access-5sssd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.993018 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:25 crc kubenswrapper[4706]: I1208 19:41:25.995764 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-kube-api-access-5nlzj" (OuterVolumeSpecName: "kube-api-access-5nlzj") pod "50a60ccd-c961-4c0c-ae74-0b0d9c6e172a" (UID: "50a60ccd-c961-4c0c-ae74-0b0d9c6e172a"). InnerVolumeSpecName "kube-api-access-5nlzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.027289 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.029499 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-248t7" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.046414 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083300 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsccl\" (UniqueName: \"kubernetes.io/projected/a2583ff3-e358-446c-9660-db7074583d8a-kube-api-access-jsccl\") pod \"a2583ff3-e358-446c-9660-db7074583d8a\" (UID: \"a2583ff3-e358-446c-9660-db7074583d8a\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083343 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w8m9\" (UniqueName: \"kubernetes.io/projected/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-kube-api-access-7w8m9\") pod \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\" (UID: \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083500 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f81f35c-cf94-4180-ad69-3d54f8c41a35-operator-scripts\") pod \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\" (UID: \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083555 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2583ff3-e358-446c-9660-db7074583d8a-operator-scripts\") pod \"a2583ff3-e358-446c-9660-db7074583d8a\" (UID: \"a2583ff3-e358-446c-9660-db7074583d8a\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083638 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-operator-scripts\") pod \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\" (UID: \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083697 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pnxb\" (UniqueName: \"kubernetes.io/projected/3f81f35c-cf94-4180-ad69-3d54f8c41a35-kube-api-access-2pnxb\") pod \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\" (UID: \"3f81f35c-cf94-4180-ad69-3d54f8c41a35\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083718 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-operator-scripts\") pod \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\" (UID: \"fd2c1239-cb7e-4bc1-af35-75fcd1613c77\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083820 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql5vb\" (UniqueName: \"kubernetes.io/projected/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-kube-api-access-ql5vb\") pod \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\" (UID: \"6d1a2245-a7e5-4f5f-a392-da5c9aec368a\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083888 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdg99\" (UniqueName: \"kubernetes.io/projected/bb919dfd-81a4-42b1-8e30-4788113753d2-kube-api-access-cdg99\") pod \"bb919dfd-81a4-42b1-8e30-4788113753d2\" (UID: \"bb919dfd-81a4-42b1-8e30-4788113753d2\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.083923 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb919dfd-81a4-42b1-8e30-4788113753d2-operator-scripts\") pod \"bb919dfd-81a4-42b1-8e30-4788113753d2\" (UID: \"bb919dfd-81a4-42b1-8e30-4788113753d2\") " Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.086779 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6d1a2245-a7e5-4f5f-a392-da5c9aec368a" (UID: "6d1a2245-a7e5-4f5f-a392-da5c9aec368a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.087115 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fd2c1239-cb7e-4bc1-af35-75fcd1613c77" (UID: "fd2c1239-cb7e-4bc1-af35-75fcd1613c77"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.090378 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f81f35c-cf94-4180-ad69-3d54f8c41a35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f81f35c-cf94-4180-ad69-3d54f8c41a35" (UID: "3f81f35c-cf94-4180-ad69-3d54f8c41a35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.091204 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-kube-api-access-ql5vb" (OuterVolumeSpecName: "kube-api-access-ql5vb") pod "6d1a2245-a7e5-4f5f-a392-da5c9aec368a" (UID: "6d1a2245-a7e5-4f5f-a392-da5c9aec368a"). InnerVolumeSpecName "kube-api-access-ql5vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.091800 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb919dfd-81a4-42b1-8e30-4788113753d2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bb919dfd-81a4-42b1-8e30-4788113753d2" (UID: "bb919dfd-81a4-42b1-8e30-4788113753d2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.092312 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2583ff3-e358-446c-9660-db7074583d8a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a2583ff3-e358-446c-9660-db7074583d8a" (UID: "a2583ff3-e358-446c-9660-db7074583d8a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.092835 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-kube-api-access-7w8m9" (OuterVolumeSpecName: "kube-api-access-7w8m9") pod "fd2c1239-cb7e-4bc1-af35-75fcd1613c77" (UID: "fd2c1239-cb7e-4bc1-af35-75fcd1613c77"). InnerVolumeSpecName "kube-api-access-7w8m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.093408 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f81f35c-cf94-4180-ad69-3d54f8c41a35-kube-api-access-2pnxb" (OuterVolumeSpecName: "kube-api-access-2pnxb") pod "3f81f35c-cf94-4180-ad69-3d54f8c41a35" (UID: "3f81f35c-cf94-4180-ad69-3d54f8c41a35"). InnerVolumeSpecName "kube-api-access-2pnxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.096380 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb919dfd-81a4-42b1-8e30-4788113753d2-kube-api-access-cdg99" (OuterVolumeSpecName: "kube-api-access-cdg99") pod "bb919dfd-81a4-42b1-8e30-4788113753d2" (UID: "bb919dfd-81a4-42b1-8e30-4788113753d2"). InnerVolumeSpecName "kube-api-access-cdg99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.098603 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2583ff3-e358-446c-9660-db7074583d8a-kube-api-access-jsccl" (OuterVolumeSpecName: "kube-api-access-jsccl") pod "a2583ff3-e358-446c-9660-db7074583d8a" (UID: "a2583ff3-e358-446c-9660-db7074583d8a"). InnerVolumeSpecName "kube-api-access-jsccl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108061 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f81f35c-cf94-4180-ad69-3d54f8c41a35-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108100 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2583ff3-e358-446c-9660-db7074583d8a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108113 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53d89be3-c6e5-4f3e-b788-4811b47ba332-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108125 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h475g\" (UniqueName: \"kubernetes.io/projected/5a7451c2-8247-406f-9d8f-eb5690a16555-kube-api-access-h475g\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108138 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108148 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108157 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pnxb\" (UniqueName: \"kubernetes.io/projected/3f81f35c-cf94-4180-ad69-3d54f8c41a35-kube-api-access-2pnxb\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108197 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108209 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sssd\" (UniqueName: \"kubernetes.io/projected/d1ae1b42-40e1-4593-9068-4f0319897bd4-kube-api-access-5sssd\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108218 4706 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108228 4706 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108239 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7451c2-8247-406f-9d8f-eb5690a16555-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108250 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trbwj\" (UniqueName: \"kubernetes.io/projected/53d89be3-c6e5-4f3e-b788-4811b47ba332-kube-api-access-trbwj\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108275 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1ae1b42-40e1-4593-9068-4f0319897bd4-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108288 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql5vb\" (UniqueName: \"kubernetes.io/projected/6d1a2245-a7e5-4f5f-a392-da5c9aec368a-kube-api-access-ql5vb\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108299 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nlzj\" (UniqueName: \"kubernetes.io/projected/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a-kube-api-access-5nlzj\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108311 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdg99\" (UniqueName: \"kubernetes.io/projected/bb919dfd-81a4-42b1-8e30-4788113753d2-kube-api-access-cdg99\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108323 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb919dfd-81a4-42b1-8e30-4788113753d2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108336 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsccl\" (UniqueName: \"kubernetes.io/projected/a2583ff3-e358-446c-9660-db7074583d8a-kube-api-access-jsccl\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108347 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w8m9\" (UniqueName: \"kubernetes.io/projected/fd2c1239-cb7e-4bc1-af35-75fcd1613c77-kube-api-access-7w8m9\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.108355 4706 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1ae1b42-40e1-4593-9068-4f0319897bd4-var-run\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.162420 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8s7bd"] Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.618281 4706 generic.go:334] "Generic (PLEG): container finished" podID="a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" containerID="a3dc3350128b2066c05ed3e4109abe891d8d32b47f84bc2a2eb2876628f0c10d" exitCode=0 Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.618616 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-thlcc" event={"ID":"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c","Type":"ContainerDied","Data":"a3dc3350128b2066c05ed3e4109abe891d8d32b47f84bc2a2eb2876628f0c10d"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.622557 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-248t7" event={"ID":"3f81f35c-cf94-4180-ad69-3d54f8c41a35","Type":"ContainerDied","Data":"bcd44ec47757503b1f466666c79295b22283a2a3d65f6ca391d57c64235f1eb8"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.622612 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcd44ec47757503b1f466666c79295b22283a2a3d65f6ca391d57c64235f1eb8" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.622705 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-248t7" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.626141 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78wps-config-hl42b" event={"ID":"d1ae1b42-40e1-4593-9068-4f0319897bd4","Type":"ContainerDied","Data":"f6ec4cde8d20a962d96c45ec71746d8a219353e2833f5ee46bed4ae871859b12"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.626196 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78wps-config-hl42b" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.626201 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6ec4cde8d20a962d96c45ec71746d8a219353e2833f5ee46bed4ae871859b12" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.632168 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2e9f-account-create-update-9kkfk" event={"ID":"fd2c1239-cb7e-4bc1-af35-75fcd1613c77","Type":"ContainerDied","Data":"1c1123a563c8838fb69c2ce676979791184c7836ee37129759387fe5d1a2d47c"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.632293 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c1123a563c8838fb69c2ce676979791184c7836ee37129759387fe5d1a2d47c" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.632413 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2e9f-account-create-update-9kkfk" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.635954 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pjffv" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.635957 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pjffv" event={"ID":"bb919dfd-81a4-42b1-8e30-4788113753d2","Type":"ContainerDied","Data":"5bf0adf855b4bd3600f2731765ef469fb3738c5bcdb02d3dc92bd888576fea3a"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.637272 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bf0adf855b4bd3600f2731765ef469fb3738c5bcdb02d3dc92bd888576fea3a" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.641630 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b7de-account-create-update-ntz9l" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.641660 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b7de-account-create-update-ntz9l" event={"ID":"5a7451c2-8247-406f-9d8f-eb5690a16555","Type":"ContainerDied","Data":"8d5042e8a09a720b789315b7bd661ad108c4fee5344eb081e02d9eaa24611707"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.641717 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d5042e8a09a720b789315b7bd661ad108c4fee5344eb081e02d9eaa24611707" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.647995 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-6lf8v" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.648396 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-6lf8v" event={"ID":"50a60ccd-c961-4c0c-ae74-0b0d9c6e172a","Type":"ContainerDied","Data":"706f5fb94216b1b0f593b8b41d8f3c7f65b88628528f2e2fb7264533690f638f"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.648456 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="706f5fb94216b1b0f593b8b41d8f3c7f65b88628528f2e2fb7264533690f638f" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.651706 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dfae-account-create-update-dhrlq" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.653084 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dfae-account-create-update-dhrlq" event={"ID":"53d89be3-c6e5-4f3e-b788-4811b47ba332","Type":"ContainerDied","Data":"34c9234c22ac0a523ff5403fe57780eb967337c285703437fb4a35d898e017e1"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.653198 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34c9234c22ac0a523ff5403fe57780eb967337c285703437fb4a35d898e017e1" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.659874 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xlsrb" event={"ID":"6507d8ce-f566-46d6-b55d-abbaccc71857","Type":"ContainerDied","Data":"277a7137bce66705b07fdf26feed278d8780b499d0b8ef8c6c60c1d43a8508c9"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.659908 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xlsrb" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.659914 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="277a7137bce66705b07fdf26feed278d8780b499d0b8ef8c6c60c1d43a8508c9" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.662686 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" event={"ID":"a2583ff3-e358-446c-9660-db7074583d8a","Type":"ContainerDied","Data":"5cd2f220ebfc90e073ad3511d3758d1ceb9149e44e8b3974bf5b5ec04cbf4d91"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.662733 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-3e4b-account-create-update-ghzkt" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.662732 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cd2f220ebfc90e073ad3511d3758d1ceb9149e44e8b3974bf5b5ec04cbf4d91" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.667977 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-04cd-account-create-update-zbfff" event={"ID":"6d1a2245-a7e5-4f5f-a392-da5c9aec368a","Type":"ContainerDied","Data":"36b7a3cf54674937b1fdb1075a42076156d95f4f8f42b722857eb3de86488575"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.668026 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36b7a3cf54674937b1fdb1075a42076156d95f4f8f42b722857eb3de86488575" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.668160 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-04cd-account-create-update-zbfff" Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.673668 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8s7bd" event={"ID":"e18991e1-84d5-4e32-9a05-ad08f31cbd60","Type":"ContainerStarted","Data":"7178f1388f4e200919070f5139b9bb0c4e4df82be32c11c9635aeac6c4da3a17"} Dec 08 19:41:26 crc kubenswrapper[4706]: I1208 19:41:26.673696 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cpmsk" Dec 08 19:41:27 crc kubenswrapper[4706]: I1208 19:41:27.160222 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-78wps-config-hl42b"] Dec 08 19:41:27 crc kubenswrapper[4706]: I1208 19:41:27.188436 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-78wps-config-hl42b"] Dec 08 19:41:27 crc kubenswrapper[4706]: I1208 19:41:27.628952 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ae1b42-40e1-4593-9068-4f0319897bd4" path="/var/lib/kubelet/pods/d1ae1b42-40e1-4593-9068-4f0319897bd4/volumes" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.138703 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.264801 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-combined-ca-bundle\") pod \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.265052 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-scripts\") pod \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.265091 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-dispersionconf\") pod \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.265148 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-etc-swift\") pod \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.265177 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-ring-data-devices\") pod \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.265226 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhn7s\" (UniqueName: \"kubernetes.io/projected/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-kube-api-access-zhn7s\") pod \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.265309 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-swiftconf\") pod \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\" (UID: \"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c\") " Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.266891 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" (UID: "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.267293 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" (UID: "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.267379 4706 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.274456 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-kube-api-access-zhn7s" (OuterVolumeSpecName: "kube-api-access-zhn7s") pod "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" (UID: "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c"). InnerVolumeSpecName "kube-api-access-zhn7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.289143 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" (UID: "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.296687 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-scripts" (OuterVolumeSpecName: "scripts") pod "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" (UID: "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.304625 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" (UID: "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.306911 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" (UID: "a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.369277 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhn7s\" (UniqueName: \"kubernetes.io/projected/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-kube-api-access-zhn7s\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.369319 4706 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.369332 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.369344 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.369356 4706 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.369368 4706 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.731403 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-thlcc" event={"ID":"a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c","Type":"ContainerDied","Data":"d5d871c568e64473eb7f698d05003e39532b2043983a99a477558022e051f3cb"} Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.731828 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5d871c568e64473eb7f698d05003e39532b2043983a99a477558022e051f3cb" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.731430 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-thlcc" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.869473 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-g8g9r"] Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870296 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f81f35c-cf94-4180-ad69-3d54f8c41a35" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870317 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f81f35c-cf94-4180-ad69-3d54f8c41a35" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870327 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb919dfd-81a4-42b1-8e30-4788113753d2" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870335 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb919dfd-81a4-42b1-8e30-4788113753d2" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870356 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1bc7a6a-8c78-4597-ba47-23e0270264e4" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870363 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1bc7a6a-8c78-4597-ba47-23e0270264e4" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870379 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2c1239-cb7e-4bc1-af35-75fcd1613c77" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870385 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2c1239-cb7e-4bc1-af35-75fcd1613c77" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870395 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6507d8ce-f566-46d6-b55d-abbaccc71857" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870401 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6507d8ce-f566-46d6-b55d-abbaccc71857" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870411 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a60ccd-c961-4c0c-ae74-0b0d9c6e172a" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870418 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a60ccd-c961-4c0c-ae74-0b0d9c6e172a" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870428 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2583ff3-e358-446c-9660-db7074583d8a" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870434 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2583ff3-e358-446c-9660-db7074583d8a" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870444 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d89be3-c6e5-4f3e-b788-4811b47ba332" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870450 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d89be3-c6e5-4f3e-b788-4811b47ba332" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870463 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a7451c2-8247-406f-9d8f-eb5690a16555" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870478 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a7451c2-8247-406f-9d8f-eb5690a16555" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870489 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" containerName="swift-ring-rebalance" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870496 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" containerName="swift-ring-rebalance" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870505 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ae1b42-40e1-4593-9068-4f0319897bd4" containerName="ovn-config" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870513 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ae1b42-40e1-4593-9068-4f0319897bd4" containerName="ovn-config" Dec 08 19:41:28 crc kubenswrapper[4706]: E1208 19:41:28.870525 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1a2245-a7e5-4f5f-a392-da5c9aec368a" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870533 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1a2245-a7e5-4f5f-a392-da5c9aec368a" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870750 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f81f35c-cf94-4180-ad69-3d54f8c41a35" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870763 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="6507d8ce-f566-46d6-b55d-abbaccc71857" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870770 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ae1b42-40e1-4593-9068-4f0319897bd4" containerName="ovn-config" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870780 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2583ff3-e358-446c-9660-db7074583d8a" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870790 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1a2245-a7e5-4f5f-a392-da5c9aec368a" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870807 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c" containerName="swift-ring-rebalance" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870817 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1bc7a6a-8c78-4597-ba47-23e0270264e4" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870824 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a60ccd-c961-4c0c-ae74-0b0d9c6e172a" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870834 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d89be3-c6e5-4f3e-b788-4811b47ba332" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870841 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb919dfd-81a4-42b1-8e30-4788113753d2" containerName="mariadb-database-create" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870875 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a7451c2-8247-406f-9d8f-eb5690a16555" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.870883 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd2c1239-cb7e-4bc1-af35-75fcd1613c77" containerName="mariadb-account-create-update" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.871765 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.877677 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.877833 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.877911 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.878275 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wbnw9" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.888675 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-g8g9r"] Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.986806 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csvxb\" (UniqueName: \"kubernetes.io/projected/3b499661-3a56-4b32-ada0-d0a132752928-kube-api-access-csvxb\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.986894 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-config-data\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:28 crc kubenswrapper[4706]: I1208 19:41:28.986946 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-combined-ca-bundle\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:29 crc kubenswrapper[4706]: I1208 19:41:29.089351 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csvxb\" (UniqueName: \"kubernetes.io/projected/3b499661-3a56-4b32-ada0-d0a132752928-kube-api-access-csvxb\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:29 crc kubenswrapper[4706]: I1208 19:41:29.089450 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-config-data\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:29 crc kubenswrapper[4706]: I1208 19:41:29.089494 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-combined-ca-bundle\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:29 crc kubenswrapper[4706]: I1208 19:41:29.109580 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-combined-ca-bundle\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:29 crc kubenswrapper[4706]: I1208 19:41:29.120990 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csvxb\" (UniqueName: \"kubernetes.io/projected/3b499661-3a56-4b32-ada0-d0a132752928-kube-api-access-csvxb\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:29 crc kubenswrapper[4706]: I1208 19:41:29.125088 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-config-data\") pod \"keystone-db-sync-g8g9r\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:29 crc kubenswrapper[4706]: I1208 19:41:29.258594 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:29 crc kubenswrapper[4706]: I1208 19:41:29.855213 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-g8g9r"] Dec 08 19:41:30 crc kubenswrapper[4706]: I1208 19:41:30.754802 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g8g9r" event={"ID":"3b499661-3a56-4b32-ada0-d0a132752928","Type":"ContainerStarted","Data":"9257a129072ac1ce734ae9a97f54a80e671c89a2c86510a6c47d8723a5718dcc"} Dec 08 19:41:32 crc kubenswrapper[4706]: I1208 19:41:32.060449 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:32 crc kubenswrapper[4706]: I1208 19:41:32.064011 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:32 crc kubenswrapper[4706]: I1208 19:41:32.780904 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:33 crc kubenswrapper[4706]: I1208 19:41:33.895926 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="9a912025-8f8e-4d61-861e-c933dafe4660" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 08 19:41:33 crc kubenswrapper[4706]: I1208 19:41:33.925185 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:33 crc kubenswrapper[4706]: I1208 19:41:33.946938 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d81adfd7-689a-496a-b4bf-c35450b63ca9-etc-swift\") pod \"swift-storage-0\" (UID: \"d81adfd7-689a-496a-b4bf-c35450b63ca9\") " pod="openstack/swift-storage-0" Dec 08 19:41:34 crc kubenswrapper[4706]: I1208 19:41:34.121543 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 08 19:41:35 crc kubenswrapper[4706]: I1208 19:41:35.454106 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:41:35 crc kubenswrapper[4706]: I1208 19:41:35.454859 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="prometheus" containerID="cri-o://5550e29689eec464a155d46fa108579fce71523aade86b30467ed46732e8ff81" gracePeriod=600 Dec 08 19:41:35 crc kubenswrapper[4706]: I1208 19:41:35.455551 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="thanos-sidecar" containerID="cri-o://22c94ccf73de5e86f0282d36e79d17564319da67bf2ad260d8fdbb3dee34a26d" gracePeriod=600 Dec 08 19:41:35 crc kubenswrapper[4706]: I1208 19:41:35.455622 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="config-reloader" containerID="cri-o://42582ba53546784a04ba0ac07e86af8166caf6957c42a3d67bd4cfa7e3bad73f" gracePeriod=600 Dec 08 19:41:36 crc kubenswrapper[4706]: I1208 19:41:36.824983 4706 generic.go:334] "Generic (PLEG): container finished" podID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerID="22c94ccf73de5e86f0282d36e79d17564319da67bf2ad260d8fdbb3dee34a26d" exitCode=0 Dec 08 19:41:36 crc kubenswrapper[4706]: I1208 19:41:36.825581 4706 generic.go:334] "Generic (PLEG): container finished" podID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerID="42582ba53546784a04ba0ac07e86af8166caf6957c42a3d67bd4cfa7e3bad73f" exitCode=0 Dec 08 19:41:36 crc kubenswrapper[4706]: I1208 19:41:36.825097 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerDied","Data":"22c94ccf73de5e86f0282d36e79d17564319da67bf2ad260d8fdbb3dee34a26d"} Dec 08 19:41:36 crc kubenswrapper[4706]: I1208 19:41:36.825643 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerDied","Data":"42582ba53546784a04ba0ac07e86af8166caf6957c42a3d67bd4cfa7e3bad73f"} Dec 08 19:41:36 crc kubenswrapper[4706]: I1208 19:41:36.825659 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerDied","Data":"5550e29689eec464a155d46fa108579fce71523aade86b30467ed46732e8ff81"} Dec 08 19:41:36 crc kubenswrapper[4706]: I1208 19:41:36.825599 4706 generic.go:334] "Generic (PLEG): container finished" podID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerID="5550e29689eec464a155d46fa108579fce71523aade86b30467ed46732e8ff81" exitCode=0 Dec 08 19:41:37 crc kubenswrapper[4706]: I1208 19:41:37.060778 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.112:9090/-/ready\": dial tcp 10.217.0.112:9090: connect: connection refused" Dec 08 19:41:42 crc kubenswrapper[4706]: I1208 19:41:42.060998 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.112:9090/-/ready\": dial tcp 10.217.0.112:9090: connect: connection refused" Dec 08 19:41:43 crc kubenswrapper[4706]: I1208 19:41:43.917830 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 08 19:41:45 crc kubenswrapper[4706]: E1208 19:41:45.010396 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 08 19:41:45 crc kubenswrapper[4706]: E1208 19:41:45.010690 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9fv45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-8s7bd_openstack(e18991e1-84d5-4e32-9a05-ad08f31cbd60): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:41:45 crc kubenswrapper[4706]: E1208 19:41:45.011985 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-8s7bd" podUID="e18991e1-84d5-4e32-9a05-ad08f31cbd60" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.756465 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.853515 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"4853c157-69b5-45ec-8fad-9e5ddea58cea\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.853582 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-config\") pod \"4853c157-69b5-45ec-8fad-9e5ddea58cea\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.853626 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp9jl\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-kube-api-access-zp9jl\") pod \"4853c157-69b5-45ec-8fad-9e5ddea58cea\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.853783 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-thanos-prometheus-http-client-file\") pod \"4853c157-69b5-45ec-8fad-9e5ddea58cea\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.853907 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4853c157-69b5-45ec-8fad-9e5ddea58cea-prometheus-metric-storage-rulefiles-0\") pod \"4853c157-69b5-45ec-8fad-9e5ddea58cea\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.853953 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4853c157-69b5-45ec-8fad-9e5ddea58cea-config-out\") pod \"4853c157-69b5-45ec-8fad-9e5ddea58cea\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.854035 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-web-config\") pod \"4853c157-69b5-45ec-8fad-9e5ddea58cea\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.854185 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-tls-assets\") pod \"4853c157-69b5-45ec-8fad-9e5ddea58cea\" (UID: \"4853c157-69b5-45ec-8fad-9e5ddea58cea\") " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.854527 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4853c157-69b5-45ec-8fad-9e5ddea58cea-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "4853c157-69b5-45ec-8fad-9e5ddea58cea" (UID: "4853c157-69b5-45ec-8fad-9e5ddea58cea"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.860718 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4853c157-69b5-45ec-8fad-9e5ddea58cea-config-out" (OuterVolumeSpecName: "config-out") pod "4853c157-69b5-45ec-8fad-9e5ddea58cea" (UID: "4853c157-69b5-45ec-8fad-9e5ddea58cea"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.860831 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "4853c157-69b5-45ec-8fad-9e5ddea58cea" (UID: "4853c157-69b5-45ec-8fad-9e5ddea58cea"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.860912 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-config" (OuterVolumeSpecName: "config") pod "4853c157-69b5-45ec-8fad-9e5ddea58cea" (UID: "4853c157-69b5-45ec-8fad-9e5ddea58cea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.861699 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "4853c157-69b5-45ec-8fad-9e5ddea58cea" (UID: "4853c157-69b5-45ec-8fad-9e5ddea58cea"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.862677 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-kube-api-access-zp9jl" (OuterVolumeSpecName: "kube-api-access-zp9jl") pod "4853c157-69b5-45ec-8fad-9e5ddea58cea" (UID: "4853c157-69b5-45ec-8fad-9e5ddea58cea"). InnerVolumeSpecName "kube-api-access-zp9jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.879218 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "4853c157-69b5-45ec-8fad-9e5ddea58cea" (UID: "4853c157-69b5-45ec-8fad-9e5ddea58cea"). InnerVolumeSpecName "pvc-305e2447-4ebb-423c-808e-09ed68f3566f". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.883374 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-web-config" (OuterVolumeSpecName: "web-config") pod "4853c157-69b5-45ec-8fad-9e5ddea58cea" (UID: "4853c157-69b5-45ec-8fad-9e5ddea58cea"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.956663 4706 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-web-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.956704 4706 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.956753 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") on node \"crc\" " Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.956766 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.956777 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp9jl\" (UniqueName: \"kubernetes.io/projected/4853c157-69b5-45ec-8fad-9e5ddea58cea-kube-api-access-zp9jl\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.956810 4706 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4853c157-69b5-45ec-8fad-9e5ddea58cea-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.956820 4706 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4853c157-69b5-45ec-8fad-9e5ddea58cea-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.956832 4706 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4853c157-69b5-45ec-8fad-9e5ddea58cea-config-out\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.964176 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g8g9r" event={"ID":"3b499661-3a56-4b32-ada0-d0a132752928","Type":"ContainerStarted","Data":"052441d1f5ffa04ff8fa2acadb9a0fb45905f41303dfb139d97ec38a230b3b82"} Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.970465 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4853c157-69b5-45ec-8fad-9e5ddea58cea","Type":"ContainerDied","Data":"8e4155b204c3112d62364734501d944668cbef0cd75413512668b584761b6e12"} Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.970598 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.970611 4706 scope.go:117] "RemoveContainer" containerID="22c94ccf73de5e86f0282d36e79d17564319da67bf2ad260d8fdbb3dee34a26d" Dec 08 19:41:45 crc kubenswrapper[4706]: E1208 19:41:45.976152 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-8s7bd" podUID="e18991e1-84d5-4e32-9a05-ad08f31cbd60" Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.988493 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 08 19:41:45 crc kubenswrapper[4706]: I1208 19:41:45.988729 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-305e2447-4ebb-423c-808e-09ed68f3566f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f") on node "crc" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.014709 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-g8g9r" podStartSLOduration=2.636659268 podStartE2EDuration="18.014679652s" podCreationTimestamp="2025-12-08 19:41:28 +0000 UTC" firstStartedPulling="2025-12-08 19:41:29.884555749 +0000 UTC m=+1172.526756752" lastFinishedPulling="2025-12-08 19:41:45.262576133 +0000 UTC m=+1187.904777136" observedRunningTime="2025-12-08 19:41:45.992782735 +0000 UTC m=+1188.634983758" watchObservedRunningTime="2025-12-08 19:41:46.014679652 +0000 UTC m=+1188.656880655" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.018376 4706 scope.go:117] "RemoveContainer" containerID="42582ba53546784a04ba0ac07e86af8166caf6957c42a3d67bd4cfa7e3bad73f" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.059126 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.089160 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.101351 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.127341 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:41:46 crc kubenswrapper[4706]: E1208 19:41:46.127808 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="thanos-sidecar" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.127833 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="thanos-sidecar" Dec 08 19:41:46 crc kubenswrapper[4706]: E1208 19:41:46.127845 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="config-reloader" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.127852 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="config-reloader" Dec 08 19:41:46 crc kubenswrapper[4706]: E1208 19:41:46.127865 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="init-config-reloader" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.127872 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="init-config-reloader" Dec 08 19:41:46 crc kubenswrapper[4706]: E1208 19:41:46.127897 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="prometheus" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.127915 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="prometheus" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.128095 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="prometheus" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.128106 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="config-reloader" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.128124 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" containerName="thanos-sidecar" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.129939 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.132809 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.136204 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.137224 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.137417 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-8tk42" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.137714 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.139074 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.145472 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.158700 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.159076 4706 scope.go:117] "RemoveContainer" containerID="5550e29689eec464a155d46fa108579fce71523aade86b30467ed46732e8ff81" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.159488 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.161219 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.161349 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eeb4b596-a4eb-4e91-bee9-123f57198cd3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.161402 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.161465 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eeb4b596-a4eb-4e91-bee9-123f57198cd3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.161488 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-config\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.161543 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eeb4b596-a4eb-4e91-bee9-123f57198cd3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.245501 4706 scope.go:117] "RemoveContainer" containerID="e2c9357f543311893043863bba13f3ac1a2ce573ff909d0205d3d74014d39082" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.264625 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.264709 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd5lb\" (UniqueName: \"kubernetes.io/projected/eeb4b596-a4eb-4e91-bee9-123f57198cd3-kube-api-access-xd5lb\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.264803 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eeb4b596-a4eb-4e91-bee9-123f57198cd3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.264914 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-config\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.265143 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eeb4b596-a4eb-4e91-bee9-123f57198cd3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.265214 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.265426 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.265471 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.265506 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.265610 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eeb4b596-a4eb-4e91-bee9-123f57198cd3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.266095 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.269122 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eeb4b596-a4eb-4e91-bee9-123f57198cd3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.272470 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.277049 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-config\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.278306 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eeb4b596-a4eb-4e91-bee9-123f57198cd3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.288721 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.288782 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e53f9d22d9359bccec8cb0090060887d02148bfeb3788e3e61a1431ce860cbd0/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.288794 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eeb4b596-a4eb-4e91-bee9-123f57198cd3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.335835 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-305e2447-4ebb-423c-808e-09ed68f3566f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-305e2447-4ebb-423c-808e-09ed68f3566f\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.368565 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.368626 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd5lb\" (UniqueName: \"kubernetes.io/projected/eeb4b596-a4eb-4e91-bee9-123f57198cd3-kube-api-access-xd5lb\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.369177 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.369881 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.369914 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.374051 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.374813 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.375582 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.375681 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eeb4b596-a4eb-4e91-bee9-123f57198cd3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.392276 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd5lb\" (UniqueName: \"kubernetes.io/projected/eeb4b596-a4eb-4e91-bee9-123f57198cd3-kube-api-access-xd5lb\") pod \"prometheus-metric-storage-0\" (UID: \"eeb4b596-a4eb-4e91-bee9-123f57198cd3\") " pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.539410 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 08 19:41:46 crc kubenswrapper[4706]: I1208 19:41:46.991154 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"2091137cb53f4b37b657f50fc0c57bc4d35cf66700f1658ad0b9e68af87c0d16"} Dec 08 19:41:47 crc kubenswrapper[4706]: I1208 19:41:47.060361 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 08 19:41:47 crc kubenswrapper[4706]: W1208 19:41:47.075197 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeeb4b596_a4eb_4e91_bee9_123f57198cd3.slice/crio-08bddf048ad3db70445618af985e8df58ea7df9674b6a65dca8fb300213c014d WatchSource:0}: Error finding container 08bddf048ad3db70445618af985e8df58ea7df9674b6a65dca8fb300213c014d: Status 404 returned error can't find the container with id 08bddf048ad3db70445618af985e8df58ea7df9674b6a65dca8fb300213c014d Dec 08 19:41:47 crc kubenswrapper[4706]: I1208 19:41:47.628050 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4853c157-69b5-45ec-8fad-9e5ddea58cea" path="/var/lib/kubelet/pods/4853c157-69b5-45ec-8fad-9e5ddea58cea/volumes" Dec 08 19:41:48 crc kubenswrapper[4706]: I1208 19:41:48.015844 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eeb4b596-a4eb-4e91-bee9-123f57198cd3","Type":"ContainerStarted","Data":"08bddf048ad3db70445618af985e8df58ea7df9674b6a65dca8fb300213c014d"} Dec 08 19:41:49 crc kubenswrapper[4706]: I1208 19:41:49.028796 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"abb17ca49e4fe69a5b7653bf771b2467482db1b0663d9b1f8656da7f595c6d47"} Dec 08 19:41:49 crc kubenswrapper[4706]: I1208 19:41:49.029334 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"c1a31ae1b39465df2dfcf11459a35333aa6cc05c7152518f3daf680a3595e487"} Dec 08 19:41:49 crc kubenswrapper[4706]: I1208 19:41:49.029347 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"06f2c09a673fd891737f8adc0d1fe08cfcb198a1eefaf5eb6cf2ce4a4859ae45"} Dec 08 19:41:49 crc kubenswrapper[4706]: I1208 19:41:49.029356 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"7325fde11464aa238103c1b8749a608a73b931ecc21671e762c626ea6bd8ae1c"} Dec 08 19:41:50 crc kubenswrapper[4706]: I1208 19:41:50.045161 4706 generic.go:334] "Generic (PLEG): container finished" podID="3b499661-3a56-4b32-ada0-d0a132752928" containerID="052441d1f5ffa04ff8fa2acadb9a0fb45905f41303dfb139d97ec38a230b3b82" exitCode=0 Dec 08 19:41:50 crc kubenswrapper[4706]: I1208 19:41:50.045277 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g8g9r" event={"ID":"3b499661-3a56-4b32-ada0-d0a132752928","Type":"ContainerDied","Data":"052441d1f5ffa04ff8fa2acadb9a0fb45905f41303dfb139d97ec38a230b3b82"} Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.057850 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eeb4b596-a4eb-4e91-bee9-123f57198cd3","Type":"ContainerStarted","Data":"5f08b7638b82fb6eac719eb1d4b5ddbd4da7364f1791de41520f30930af03909"} Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.062776 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"d86a4c63cbfbb37cb524359360298df940cc1d84fe4e3322fe7d624fd8eedead"} Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.062818 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"1b55f6662aa43c30c96ee14ee7774c02171991697d50fee8fdab551121adfe17"} Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.062830 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"c86be7aefe2206ae1578f6b390e046fdd190a0c1ca920b2cb6ac37666dc7fa5c"} Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.471702 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.601060 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csvxb\" (UniqueName: \"kubernetes.io/projected/3b499661-3a56-4b32-ada0-d0a132752928-kube-api-access-csvxb\") pod \"3b499661-3a56-4b32-ada0-d0a132752928\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.601153 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-config-data\") pod \"3b499661-3a56-4b32-ada0-d0a132752928\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.601179 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-combined-ca-bundle\") pod \"3b499661-3a56-4b32-ada0-d0a132752928\" (UID: \"3b499661-3a56-4b32-ada0-d0a132752928\") " Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.636627 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b499661-3a56-4b32-ada0-d0a132752928-kube-api-access-csvxb" (OuterVolumeSpecName: "kube-api-access-csvxb") pod "3b499661-3a56-4b32-ada0-d0a132752928" (UID: "3b499661-3a56-4b32-ada0-d0a132752928"). InnerVolumeSpecName "kube-api-access-csvxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.644745 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b499661-3a56-4b32-ada0-d0a132752928" (UID: "3b499661-3a56-4b32-ada0-d0a132752928"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.704187 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csvxb\" (UniqueName: \"kubernetes.io/projected/3b499661-3a56-4b32-ada0-d0a132752928-kube-api-access-csvxb\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.704229 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.722504 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-config-data" (OuterVolumeSpecName: "config-data") pod "3b499661-3a56-4b32-ada0-d0a132752928" (UID: "3b499661-3a56-4b32-ada0-d0a132752928"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:41:51 crc kubenswrapper[4706]: I1208 19:41:51.805818 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b499661-3a56-4b32-ada0-d0a132752928-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.100041 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"a2d84425ac1351f73b257e6c752287f3305c4927165063fe0021f9a5793889d6"} Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.103533 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g8g9r" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.104167 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g8g9r" event={"ID":"3b499661-3a56-4b32-ada0-d0a132752928","Type":"ContainerDied","Data":"9257a129072ac1ce734ae9a97f54a80e671c89a2c86510a6c47d8723a5718dcc"} Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.104197 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9257a129072ac1ce734ae9a97f54a80e671c89a2c86510a6c47d8723a5718dcc" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.370049 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-979rj"] Dec 08 19:41:52 crc kubenswrapper[4706]: E1208 19:41:52.370892 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b499661-3a56-4b32-ada0-d0a132752928" containerName="keystone-db-sync" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.370915 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b499661-3a56-4b32-ada0-d0a132752928" containerName="keystone-db-sync" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.371139 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b499661-3a56-4b32-ada0-d0a132752928" containerName="keystone-db-sync" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.378282 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.395424 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rrxz6"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.397202 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.401802 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.402009 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wbnw9" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.402126 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.402386 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.402510 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.429351 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-979rj"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.489722 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rrxz6"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.536799 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.536871 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-config-data\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.536915 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwdjx\" (UniqueName: \"kubernetes.io/projected/8d699fdc-91ac-446c-812c-2dd9f5045947-kube-api-access-vwdjx\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.536966 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.537000 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-combined-ca-bundle\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.537037 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-credential-keys\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.537071 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-dns-svc\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.537104 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdx9q\" (UniqueName: \"kubernetes.io/projected/3ee2aad1-63b7-43fc-9615-0c7263f86815-kube-api-access-rdx9q\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.537149 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-scripts\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.537206 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-config\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.537247 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-fernet-keys\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.616249 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-dkmkk"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.617845 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.627128 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.627353 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.629147 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8p7hk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.638859 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-fernet-keys\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.638924 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.638961 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-config-data\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.638994 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-config\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639013 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj2f5\" (UniqueName: \"kubernetes.io/projected/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-kube-api-access-sj2f5\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639040 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwdjx\" (UniqueName: \"kubernetes.io/projected/8d699fdc-91ac-446c-812c-2dd9f5045947-kube-api-access-vwdjx\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639073 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-combined-ca-bundle\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639099 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639127 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-combined-ca-bundle\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639152 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-credential-keys\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639180 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-dns-svc\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639210 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdx9q\" (UniqueName: \"kubernetes.io/projected/3ee2aad1-63b7-43fc-9615-0c7263f86815-kube-api-access-rdx9q\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639250 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-scripts\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.639319 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-config\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.640651 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-config\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.656767 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dkmkk"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.656945 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.658810 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-dns-svc\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.659211 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.683387 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-credential-keys\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.684948 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-scripts\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.691284 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-config-data\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.703798 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-combined-ca-bundle\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.705011 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-fernet-keys\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.752500 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-config\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.752584 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj2f5\" (UniqueName: \"kubernetes.io/projected/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-kube-api-access-sj2f5\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.752701 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-combined-ca-bundle\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.759985 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwdjx\" (UniqueName: \"kubernetes.io/projected/8d699fdc-91ac-446c-812c-2dd9f5045947-kube-api-access-vwdjx\") pod \"keystone-bootstrap-rrxz6\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.760064 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-combined-ca-bundle\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.776562 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.776724 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdx9q\" (UniqueName: \"kubernetes.io/projected/3ee2aad1-63b7-43fc-9615-0c7263f86815-kube-api-access-rdx9q\") pod \"dnsmasq-dns-f877ddd87-979rj\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.779167 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-config\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.810945 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj2f5\" (UniqueName: \"kubernetes.io/projected/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-kube-api-access-sj2f5\") pod \"neutron-db-sync-dkmkk\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.853884 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-979rj"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.855143 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.865292 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-75hmx"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.869687 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.881610 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-brz4k"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.884213 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.911368 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.911860 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-z4r6h" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.912216 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.912463 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-75hmx"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.920645 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-brz4k"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.977575 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-config\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.977638 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.977850 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmz2p\" (UniqueName: \"kubernetes.io/projected/6e850531-c236-429d-abdc-ab69df9a82a5-kube-api-access-xmz2p\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.978009 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.978335 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.987234 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ff9kc"] Dec 08 19:41:52 crc kubenswrapper[4706]: I1208 19:41:52.989288 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.015862 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-c4h7r" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.016149 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.027248 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-lhc9c"] Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.029234 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.034851 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.037464 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.039138 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.040220 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bcpc2" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.058598 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-59fls"] Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.060733 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.077805 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.078217 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.078511 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.079012 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-fzwvl" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082178 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78d8k\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-kube-api-access-78d8k\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082225 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-config\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082247 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082282 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-db-sync-config-data\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082311 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-config-data\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082398 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-etc-machine-id\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082839 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-combined-ca-bundle\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082884 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-certs\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082909 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm24w\" (UniqueName: \"kubernetes.io/projected/d6007166-5c77-46c7-8c2a-e4ff780725d6-kube-api-access-sm24w\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082934 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmz2p\" (UniqueName: \"kubernetes.io/projected/6e850531-c236-429d-abdc-ab69df9a82a5-kube-api-access-xmz2p\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082958 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6s9t\" (UniqueName: \"kubernetes.io/projected/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-kube-api-access-t6s9t\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.082976 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-scripts\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083003 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-scripts\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083024 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkwnh\" (UniqueName: \"kubernetes.io/projected/8677c722-159f-4509-a1b1-114817ef241b-kube-api-access-fkwnh\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083046 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083063 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-combined-ca-bundle\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083096 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-config-data\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083117 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-scripts\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083134 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-combined-ca-bundle\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083160 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-config-data\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083182 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-db-sync-config-data\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083225 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-combined-ca-bundle\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083244 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6007166-5c77-46c7-8c2a-e4ff780725d6-logs\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083288 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083547 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-config\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.083577 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.084173 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.084369 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.109739 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ff9kc"] Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.133383 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmz2p\" (UniqueName: \"kubernetes.io/projected/6e850531-c236-429d-abdc-ab69df9a82a5-kube-api-access-xmz2p\") pod \"dnsmasq-dns-68dcc9cf6f-75hmx\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.157665 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-59fls"] Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.175877 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.183219 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.190702 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.190960 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.191653 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208125 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6s9t\" (UniqueName: \"kubernetes.io/projected/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-kube-api-access-t6s9t\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208180 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-log-httpd\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208211 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-scripts\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208237 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-scripts\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208324 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-scripts\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208366 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkwnh\" (UniqueName: \"kubernetes.io/projected/8677c722-159f-4509-a1b1-114817ef241b-kube-api-access-fkwnh\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208393 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-combined-ca-bundle\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208481 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-config-data\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208501 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-config-data\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208525 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-scripts\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208550 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-combined-ca-bundle\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208585 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208617 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-config-data\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208635 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-db-sync-config-data\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208702 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g265b\" (UniqueName: \"kubernetes.io/projected/a298368b-7b89-4c7a-8427-330e9d16ea00-kube-api-access-g265b\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208775 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-combined-ca-bundle\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208805 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6007166-5c77-46c7-8c2a-e4ff780725d6-logs\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208851 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-run-httpd\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208899 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78d8k\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-kube-api-access-78d8k\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208940 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-db-sync-config-data\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.208994 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-config-data\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.209016 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-etc-machine-id\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.209037 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.209071 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-combined-ca-bundle\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.209123 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-certs\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.209161 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm24w\" (UniqueName: \"kubernetes.io/projected/d6007166-5c77-46c7-8c2a-e4ff780725d6-kube-api-access-sm24w\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.220907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6007166-5c77-46c7-8c2a-e4ff780725d6-logs\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.221843 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-scripts\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.222328 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-etc-machine-id\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.223671 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-scripts\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.229294 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-db-sync-config-data\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.238880 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-combined-ca-bundle\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.241782 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-scripts\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.255007 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-combined-ca-bundle\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.257716 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-db-sync-config-data\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.258469 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-config-data\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.259485 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-config-data\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.261843 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-combined-ca-bundle\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.261917 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lhc9c"] Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.262034 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-config-data\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.262994 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkwnh\" (UniqueName: \"kubernetes.io/projected/8677c722-159f-4509-a1b1-114817ef241b-kube-api-access-fkwnh\") pod \"barbican-db-sync-ff9kc\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.263404 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm24w\" (UniqueName: \"kubernetes.io/projected/d6007166-5c77-46c7-8c2a-e4ff780725d6-kube-api-access-sm24w\") pod \"placement-db-sync-lhc9c\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.263856 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78d8k\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-kube-api-access-78d8k\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.265189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-certs\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.270375 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-combined-ca-bundle\") pod \"cloudkitty-db-sync-59fls\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.273829 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6s9t\" (UniqueName: \"kubernetes.io/projected/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-kube-api-access-t6s9t\") pod \"cinder-db-sync-brz4k\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.296038 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.312673 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-brz4k" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.314362 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.314427 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-log-httpd\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.314451 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-scripts\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.314514 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-config-data\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.314536 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.315057 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g265b\" (UniqueName: \"kubernetes.io/projected/a298368b-7b89-4c7a-8427-330e9d16ea00-kube-api-access-g265b\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.315141 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-run-httpd\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.335170 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-run-httpd\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.336707 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-log-httpd\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.344460 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.350519 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-scripts\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.351484 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.358204 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-config-data\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.369127 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g265b\" (UniqueName: \"kubernetes.io/projected/a298368b-7b89-4c7a-8427-330e9d16ea00-kube-api-access-g265b\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.380185 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.486046 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.628458 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lhc9c" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.642328 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:41:53 crc kubenswrapper[4706]: I1208 19:41:53.806935 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rrxz6"] Dec 08 19:41:54 crc kubenswrapper[4706]: I1208 19:41:54.198580 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rrxz6" event={"ID":"8d699fdc-91ac-446c-812c-2dd9f5045947","Type":"ContainerStarted","Data":"48ae9e94636771901d87d2296b2e7cdf9393df96a3446da6c221e8d851ce33e2"} Dec 08 19:41:54 crc kubenswrapper[4706]: I1208 19:41:54.202500 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dkmkk"] Dec 08 19:41:54 crc kubenswrapper[4706]: I1208 19:41:54.288246 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-979rj"] Dec 08 19:41:54 crc kubenswrapper[4706]: I1208 19:41:54.324548 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"20c8492b72bebea2561b988cb9ec08615b83fe71d499da5133da2924cc669419"} Dec 08 19:41:54 crc kubenswrapper[4706]: I1208 19:41:54.702464 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-75hmx"] Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:54.995978 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-59fls"] Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.008667 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lhc9c"] Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.030363 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ff9kc"] Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.045522 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:41:58 crc kubenswrapper[4706]: W1208 19:41:55.063493 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode68efceb_8175_4600_bc86_0d9d5d9d5dbd.slice/crio-9daeb96ebafe0f18462455c507773e2b8569079f99a7b2f4372b5cc763a3a1e9 WatchSource:0}: Error finding container 9daeb96ebafe0f18462455c507773e2b8569079f99a7b2f4372b5cc763a3a1e9: Status 404 returned error can't find the container with id 9daeb96ebafe0f18462455c507773e2b8569079f99a7b2f4372b5cc763a3a1e9 Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.162566 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-brz4k"] Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.278243 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.398501 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"de0750c1bc04c5fa329600f6be32f98e8fbabab5419ed89bd1bd3841568b9a7e"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.403821 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-59fls" event={"ID":"e68efceb-8175-4600-bc86-0d9d5d9d5dbd","Type":"ContainerStarted","Data":"9daeb96ebafe0f18462455c507773e2b8569079f99a7b2f4372b5cc763a3a1e9"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.428075 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-brz4k" event={"ID":"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e","Type":"ContainerStarted","Data":"620c4841ad21c6859c2997a524fb76855866089caaf868ade99e023c4dd3ba57"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.433645 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lhc9c" event={"ID":"d6007166-5c77-46c7-8c2a-e4ff780725d6","Type":"ContainerStarted","Data":"8a86b8a328cdf3e06ff643f32cd2376d55387dc4b32f73ed547ac12d4dafd781"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.438701 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dkmkk" event={"ID":"18ef7995-14d2-4c7a-8b30-8aaac1c49b92","Type":"ContainerStarted","Data":"75ea842913c75d2f26e428dc0a0703205af3a7cea0a7ce82fee7e017f69ca031"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.438747 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dkmkk" event={"ID":"18ef7995-14d2-4c7a-8b30-8aaac1c49b92","Type":"ContainerStarted","Data":"100b5d19b5b82eb64d96177a54638bdae0dae7b127709a9d3c7a7934cdfbb76b"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.458423 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" event={"ID":"6e850531-c236-429d-abdc-ab69df9a82a5","Type":"ContainerStarted","Data":"b5428d0e273dc80baddf451fc6f0c847c444854c8048c8550e31c8d5e867cc71"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.469022 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-dkmkk" podStartSLOduration=3.468994364 podStartE2EDuration="3.468994364s" podCreationTimestamp="2025-12-08 19:41:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:55.460589953 +0000 UTC m=+1198.102790956" watchObservedRunningTime="2025-12-08 19:41:55.468994364 +0000 UTC m=+1198.111195367" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.476517 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rrxz6" event={"ID":"8d699fdc-91ac-446c-812c-2dd9f5045947","Type":"ContainerStarted","Data":"bc2a3c7a3ebbe36ae3f29c3fdd4bc71d23a486cf3882c63ecb8cfb1c56b6ab4f"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.480566 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a298368b-7b89-4c7a-8427-330e9d16ea00","Type":"ContainerStarted","Data":"4ed782f23b2b1555bd0e1ee653c5e2934bc657b4ddf6a452118ab086b974a5a6"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.488423 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ff9kc" event={"ID":"8677c722-159f-4509-a1b1-114817ef241b","Type":"ContainerStarted","Data":"b371a3f304c0de07cca76b5b6f742fd0f48623c7a76df0a4bc4428e452067606"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.541619 4706 generic.go:334] "Generic (PLEG): container finished" podID="3ee2aad1-63b7-43fc-9615-0c7263f86815" containerID="b7f56d3b659ab693ca933e0ee8b31238a57bcf099169491ed1d68f8d10184877" exitCode=0 Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.541681 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-979rj" event={"ID":"3ee2aad1-63b7-43fc-9615-0c7263f86815","Type":"ContainerDied","Data":"b7f56d3b659ab693ca933e0ee8b31238a57bcf099169491ed1d68f8d10184877"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.541717 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-979rj" event={"ID":"3ee2aad1-63b7-43fc-9615-0c7263f86815","Type":"ContainerStarted","Data":"5d1bdadc52418067f262260c5cb61006b155c28f34166f723691c66836cc09ff"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:55.696917 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rrxz6" podStartSLOduration=3.69689604 podStartE2EDuration="3.69689604s" podCreationTimestamp="2025-12-08 19:41:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:55.5477865 +0000 UTC m=+1198.189987513" watchObservedRunningTime="2025-12-08 19:41:55.69689604 +0000 UTC m=+1198.339097043" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:56.622588 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"7df5ef9f4c2a5a2940ddd6a66a417e4af2a2797f16f2b2620173f0cc74e05c36"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:56.627526 4706 generic.go:334] "Generic (PLEG): container finished" podID="6e850531-c236-429d-abdc-ab69df9a82a5" containerID="6b235e9e7854b41ffa914ee7123a00d2d71998328b98ef842b5765d14e75e2f0" exitCode=0 Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:56.628060 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" event={"ID":"6e850531-c236-429d-abdc-ab69df9a82a5","Type":"ContainerStarted","Data":"4ac3478f9c35a5e789c84e27d046cf313d7037e8d39af6c7cc8b6431c9adb174"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:56.628305 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" event={"ID":"6e850531-c236-429d-abdc-ab69df9a82a5","Type":"ContainerDied","Data":"6b235e9e7854b41ffa914ee7123a00d2d71998328b98ef842b5765d14e75e2f0"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:56.629651 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:56.665292 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" podStartSLOduration=4.665246752 podStartE2EDuration="4.665246752s" podCreationTimestamp="2025-12-08 19:41:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:41:56.651605352 +0000 UTC m=+1199.293806355" watchObservedRunningTime="2025-12-08 19:41:56.665246752 +0000 UTC m=+1199.307447755" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:57.677897 4706 generic.go:334] "Generic (PLEG): container finished" podID="eeb4b596-a4eb-4e91-bee9-123f57198cd3" containerID="5f08b7638b82fb6eac719eb1d4b5ddbd4da7364f1791de41520f30930af03909" exitCode=0 Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:57.678719 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eeb4b596-a4eb-4e91-bee9-123f57198cd3","Type":"ContainerDied","Data":"5f08b7638b82fb6eac719eb1d4b5ddbd4da7364f1791de41520f30930af03909"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:57.745190 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"25a24789ce4ce70e2138d6cf2c1158278e5e30ec3da6e0bdcdfb26f918c4bdeb"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.342673 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.475013 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-sb\") pod \"3ee2aad1-63b7-43fc-9615-0c7263f86815\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.475640 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdx9q\" (UniqueName: \"kubernetes.io/projected/3ee2aad1-63b7-43fc-9615-0c7263f86815-kube-api-access-rdx9q\") pod \"3ee2aad1-63b7-43fc-9615-0c7263f86815\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.475831 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-dns-svc\") pod \"3ee2aad1-63b7-43fc-9615-0c7263f86815\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.475941 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-config\") pod \"3ee2aad1-63b7-43fc-9615-0c7263f86815\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.475988 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-nb\") pod \"3ee2aad1-63b7-43fc-9615-0c7263f86815\" (UID: \"3ee2aad1-63b7-43fc-9615-0c7263f86815\") " Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.492442 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ee2aad1-63b7-43fc-9615-0c7263f86815-kube-api-access-rdx9q" (OuterVolumeSpecName: "kube-api-access-rdx9q") pod "3ee2aad1-63b7-43fc-9615-0c7263f86815" (UID: "3ee2aad1-63b7-43fc-9615-0c7263f86815"). InnerVolumeSpecName "kube-api-access-rdx9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.533817 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ee2aad1-63b7-43fc-9615-0c7263f86815" (UID: "3ee2aad1-63b7-43fc-9615-0c7263f86815"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.540283 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-config" (OuterVolumeSpecName: "config") pod "3ee2aad1-63b7-43fc-9615-0c7263f86815" (UID: "3ee2aad1-63b7-43fc-9615-0c7263f86815"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.577624 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ee2aad1-63b7-43fc-9615-0c7263f86815" (UID: "3ee2aad1-63b7-43fc-9615-0c7263f86815"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.585026 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.585081 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.585092 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.585104 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdx9q\" (UniqueName: \"kubernetes.io/projected/3ee2aad1-63b7-43fc-9615-0c7263f86815-kube-api-access-rdx9q\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.625146 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ee2aad1-63b7-43fc-9615-0c7263f86815" (UID: "3ee2aad1-63b7-43fc-9615-0c7263f86815"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.702823 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee2aad1-63b7-43fc-9615-0c7263f86815-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.813941 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eeb4b596-a4eb-4e91-bee9-123f57198cd3","Type":"ContainerStarted","Data":"5b318b673d2a1dd8ebc4f46d29d000cae007563ef2914b4277726e64687ff566"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.846467 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"fd6471f9b4fac289b79df5c8a503c881ade894855e63d25b6537ea9e91831fe4"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.851935 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-979rj" event={"ID":"3ee2aad1-63b7-43fc-9615-0c7263f86815","Type":"ContainerDied","Data":"5d1bdadc52418067f262260c5cb61006b155c28f34166f723691c66836cc09ff"} Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.852011 4706 scope.go:117] "RemoveContainer" containerID="b7f56d3b659ab693ca933e0ee8b31238a57bcf099169491ed1d68f8d10184877" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.852213 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-979rj" Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.956416 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-979rj"] Dec 08 19:41:58 crc kubenswrapper[4706]: I1208 19:41:58.982591 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-979rj"] Dec 08 19:41:59 crc kubenswrapper[4706]: I1208 19:41:59.655917 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ee2aad1-63b7-43fc-9615-0c7263f86815" path="/var/lib/kubelet/pods/3ee2aad1-63b7-43fc-9615-0c7263f86815/volumes" Dec 08 19:41:59 crc kubenswrapper[4706]: I1208 19:41:59.912587 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"425f0068c6b8bd10c45f38803179729b1877e2ae2fdc207342fda80b9935342d"} Dec 08 19:42:00 crc kubenswrapper[4706]: I1208 19:42:00.947969 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d81adfd7-689a-496a-b4bf-c35450b63ca9","Type":"ContainerStarted","Data":"9948a1ded3364e8a09d41a6113c86c425099252dfb50a14b149e4875489e647b"} Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.012368 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=54.547453335 podStartE2EDuration="1m1.012314035s" podCreationTimestamp="2025-12-08 19:41:00 +0000 UTC" firstStartedPulling="2025-12-08 19:41:46.194517662 +0000 UTC m=+1188.836718665" lastFinishedPulling="2025-12-08 19:41:52.659378362 +0000 UTC m=+1195.301579365" observedRunningTime="2025-12-08 19:42:01.002139714 +0000 UTC m=+1203.644340727" watchObservedRunningTime="2025-12-08 19:42:01.012314035 +0000 UTC m=+1203.654515038" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.308144 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-75hmx"] Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.308514 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" podUID="6e850531-c236-429d-abdc-ab69df9a82a5" containerName="dnsmasq-dns" containerID="cri-o://4ac3478f9c35a5e789c84e27d046cf313d7037e8d39af6c7cc8b6431c9adb174" gracePeriod=10 Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.310575 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.358295 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c86wj"] Dec 08 19:42:01 crc kubenswrapper[4706]: E1208 19:42:01.359116 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee2aad1-63b7-43fc-9615-0c7263f86815" containerName="init" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.359141 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee2aad1-63b7-43fc-9615-0c7263f86815" containerName="init" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.359508 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ee2aad1-63b7-43fc-9615-0c7263f86815" containerName="init" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.361226 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.367125 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.406523 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c86wj"] Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.499771 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.501241 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-config\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.501644 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.501806 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.501939 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.502376 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrxtw\" (UniqueName: \"kubernetes.io/projected/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-kube-api-access-qrxtw\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.605248 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.605337 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-config\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.605377 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.605413 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.605833 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.605968 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrxtw\" (UniqueName: \"kubernetes.io/projected/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-kube-api-access-qrxtw\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.606798 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-config\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.606820 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.607622 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.607754 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.608198 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.635757 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrxtw\" (UniqueName: \"kubernetes.io/projected/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-kube-api-access-qrxtw\") pod \"dnsmasq-dns-58dd9ff6bc-c86wj\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:01 crc kubenswrapper[4706]: I1208 19:42:01.686700 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.000610 4706 generic.go:334] "Generic (PLEG): container finished" podID="6e850531-c236-429d-abdc-ab69df9a82a5" containerID="4ac3478f9c35a5e789c84e27d046cf313d7037e8d39af6c7cc8b6431c9adb174" exitCode=0 Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.000758 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" event={"ID":"6e850531-c236-429d-abdc-ab69df9a82a5","Type":"ContainerDied","Data":"4ac3478f9c35a5e789c84e27d046cf313d7037e8d39af6c7cc8b6431c9adb174"} Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.255040 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c86wj"] Dec 08 19:42:02 crc kubenswrapper[4706]: W1208 19:42:02.271809 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode87e3ba7_b301_4df4_9ca5_2f3b735c622f.slice/crio-19a49a12654ad4d7464703cd6fbce9deff82c446454e90d89a6cef77fb367b0f WatchSource:0}: Error finding container 19a49a12654ad4d7464703cd6fbce9deff82c446454e90d89a6cef77fb367b0f: Status 404 returned error can't find the container with id 19a49a12654ad4d7464703cd6fbce9deff82c446454e90d89a6cef77fb367b0f Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.859406 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.970043 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-sb\") pod \"6e850531-c236-429d-abdc-ab69df9a82a5\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.970318 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmz2p\" (UniqueName: \"kubernetes.io/projected/6e850531-c236-429d-abdc-ab69df9a82a5-kube-api-access-xmz2p\") pod \"6e850531-c236-429d-abdc-ab69df9a82a5\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.970379 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-config\") pod \"6e850531-c236-429d-abdc-ab69df9a82a5\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.970854 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-nb\") pod \"6e850531-c236-429d-abdc-ab69df9a82a5\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.970924 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-dns-svc\") pod \"6e850531-c236-429d-abdc-ab69df9a82a5\" (UID: \"6e850531-c236-429d-abdc-ab69df9a82a5\") " Dec 08 19:42:02 crc kubenswrapper[4706]: I1208 19:42:02.983713 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e850531-c236-429d-abdc-ab69df9a82a5-kube-api-access-xmz2p" (OuterVolumeSpecName: "kube-api-access-xmz2p") pod "6e850531-c236-429d-abdc-ab69df9a82a5" (UID: "6e850531-c236-429d-abdc-ab69df9a82a5"). InnerVolumeSpecName "kube-api-access-xmz2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.029670 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.029809 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-75hmx" event={"ID":"6e850531-c236-429d-abdc-ab69df9a82a5","Type":"ContainerDied","Data":"b5428d0e273dc80baddf451fc6f0c847c444854c8048c8550e31c8d5e867cc71"} Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.030581 4706 scope.go:117] "RemoveContainer" containerID="4ac3478f9c35a5e789c84e27d046cf313d7037e8d39af6c7cc8b6431c9adb174" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.031542 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" event={"ID":"e87e3ba7-b301-4df4-9ca5-2f3b735c622f","Type":"ContainerStarted","Data":"19a49a12654ad4d7464703cd6fbce9deff82c446454e90d89a6cef77fb367b0f"} Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.031805 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6e850531-c236-429d-abdc-ab69df9a82a5" (UID: "6e850531-c236-429d-abdc-ab69df9a82a5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.040441 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6e850531-c236-429d-abdc-ab69df9a82a5" (UID: "6e850531-c236-429d-abdc-ab69df9a82a5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.047613 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6e850531-c236-429d-abdc-ab69df9a82a5" (UID: "6e850531-c236-429d-abdc-ab69df9a82a5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.058671 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-config" (OuterVolumeSpecName: "config") pod "6e850531-c236-429d-abdc-ab69df9a82a5" (UID: "6e850531-c236-429d-abdc-ab69df9a82a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.066477 4706 scope.go:117] "RemoveContainer" containerID="6b235e9e7854b41ffa914ee7123a00d2d71998328b98ef842b5765d14e75e2f0" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.073639 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.073677 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmz2p\" (UniqueName: \"kubernetes.io/projected/6e850531-c236-429d-abdc-ab69df9a82a5-kube-api-access-xmz2p\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.073693 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.073707 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.073717 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e850531-c236-429d-abdc-ab69df9a82a5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.369703 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-75hmx"] Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.378307 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-75hmx"] Dec 08 19:42:03 crc kubenswrapper[4706]: I1208 19:42:03.624493 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e850531-c236-429d-abdc-ab69df9a82a5" path="/var/lib/kubelet/pods/6e850531-c236-429d-abdc-ab69df9a82a5/volumes" Dec 08 19:42:04 crc kubenswrapper[4706]: I1208 19:42:04.049530 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eeb4b596-a4eb-4e91-bee9-123f57198cd3","Type":"ContainerStarted","Data":"b7958448adbfbc7601e4f27423e672a841fc60eb9e5f261dfae91edfb656536d"} Dec 08 19:42:04 crc kubenswrapper[4706]: I1208 19:42:04.049583 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eeb4b596-a4eb-4e91-bee9-123f57198cd3","Type":"ContainerStarted","Data":"b8635208b69f750859ec44f353fc099044232ecaf290cc0b585e1e9678d3b924"} Dec 08 19:42:04 crc kubenswrapper[4706]: I1208 19:42:04.054885 4706 generic.go:334] "Generic (PLEG): container finished" podID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" containerID="ddd93e30d02810dea6a7c9829ce9248c26bdd714b7214f2dcd3cefd454732f81" exitCode=0 Dec 08 19:42:04 crc kubenswrapper[4706]: I1208 19:42:04.054933 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" event={"ID":"e87e3ba7-b301-4df4-9ca5-2f3b735c622f","Type":"ContainerDied","Data":"ddd93e30d02810dea6a7c9829ce9248c26bdd714b7214f2dcd3cefd454732f81"} Dec 08 19:42:04 crc kubenswrapper[4706]: I1208 19:42:04.081906 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.081878961 podStartE2EDuration="18.081878961s" podCreationTimestamp="2025-12-08 19:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:04.07732332 +0000 UTC m=+1206.719524323" watchObservedRunningTime="2025-12-08 19:42:04.081878961 +0000 UTC m=+1206.724079954" Dec 08 19:42:05 crc kubenswrapper[4706]: I1208 19:42:05.071651 4706 generic.go:334] "Generic (PLEG): container finished" podID="8d699fdc-91ac-446c-812c-2dd9f5045947" containerID="bc2a3c7a3ebbe36ae3f29c3fdd4bc71d23a486cf3882c63ecb8cfb1c56b6ab4f" exitCode=0 Dec 08 19:42:05 crc kubenswrapper[4706]: I1208 19:42:05.071723 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rrxz6" event={"ID":"8d699fdc-91ac-446c-812c-2dd9f5045947","Type":"ContainerDied","Data":"bc2a3c7a3ebbe36ae3f29c3fdd4bc71d23a486cf3882c63ecb8cfb1c56b6ab4f"} Dec 08 19:42:05 crc kubenswrapper[4706]: I1208 19:42:05.835953 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:42:05 crc kubenswrapper[4706]: I1208 19:42:05.836295 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:42:06 crc kubenswrapper[4706]: I1208 19:42:06.541174 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 08 19:42:10 crc kubenswrapper[4706]: E1208 19:42:10.837434 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 08 19:42:10 crc kubenswrapper[4706]: E1208 19:42:10.838455 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sm24w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-lhc9c_openstack(d6007166-5c77-46c7-8c2a-e4ff780725d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:42:10 crc kubenswrapper[4706]: E1208 19:42:10.840011 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-lhc9c" podUID="d6007166-5c77-46c7-8c2a-e4ff780725d6" Dec 08 19:42:11 crc kubenswrapper[4706]: E1208 19:42:11.170554 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-lhc9c" podUID="d6007166-5c77-46c7-8c2a-e4ff780725d6" Dec 08 19:42:15 crc kubenswrapper[4706]: E1208 19:42:15.832426 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 08 19:42:15 crc kubenswrapper[4706]: E1208 19:42:15.833464 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n7bh676h6h5f7h598h689h65bhd9h6dh596h55h86h8ch5ddh584h5dch5b6h596hfh559h7dhd8h67hdbhcdh67hch94h99h549h5dh57cq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g265b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(a298368b-7b89-4c7a-8427-330e9d16ea00): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:42:16 crc kubenswrapper[4706]: I1208 19:42:16.541499 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 08 19:42:16 crc kubenswrapper[4706]: I1208 19:42:16.549306 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 08 19:42:17 crc kubenswrapper[4706]: I1208 19:42:17.262280 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 08 19:42:26 crc kubenswrapper[4706]: I1208 19:42:26.373756 4706 generic.go:334] "Generic (PLEG): container finished" podID="18ef7995-14d2-4c7a-8b30-8aaac1c49b92" containerID="75ea842913c75d2f26e428dc0a0703205af3a7cea0a7ce82fee7e017f69ca031" exitCode=0 Dec 08 19:42:26 crc kubenswrapper[4706]: I1208 19:42:26.373871 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dkmkk" event={"ID":"18ef7995-14d2-4c7a-8b30-8aaac1c49b92","Type":"ContainerDied","Data":"75ea842913c75d2f26e428dc0a0703205af3a7cea0a7ce82fee7e017f69ca031"} Dec 08 19:42:26 crc kubenswrapper[4706]: I1208 19:42:26.993674 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.079770 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-config-data\") pod \"8d699fdc-91ac-446c-812c-2dd9f5045947\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.080100 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwdjx\" (UniqueName: \"kubernetes.io/projected/8d699fdc-91ac-446c-812c-2dd9f5045947-kube-api-access-vwdjx\") pod \"8d699fdc-91ac-446c-812c-2dd9f5045947\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.081111 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-scripts\") pod \"8d699fdc-91ac-446c-812c-2dd9f5045947\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.081172 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-combined-ca-bundle\") pod \"8d699fdc-91ac-446c-812c-2dd9f5045947\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.081276 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-fernet-keys\") pod \"8d699fdc-91ac-446c-812c-2dd9f5045947\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.081332 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-credential-keys\") pod \"8d699fdc-91ac-446c-812c-2dd9f5045947\" (UID: \"8d699fdc-91ac-446c-812c-2dd9f5045947\") " Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.086925 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d699fdc-91ac-446c-812c-2dd9f5045947-kube-api-access-vwdjx" (OuterVolumeSpecName: "kube-api-access-vwdjx") pod "8d699fdc-91ac-446c-812c-2dd9f5045947" (UID: "8d699fdc-91ac-446c-812c-2dd9f5045947"). InnerVolumeSpecName "kube-api-access-vwdjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.087078 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8d699fdc-91ac-446c-812c-2dd9f5045947" (UID: "8d699fdc-91ac-446c-812c-2dd9f5045947"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.087409 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8d699fdc-91ac-446c-812c-2dd9f5045947" (UID: "8d699fdc-91ac-446c-812c-2dd9f5045947"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.089692 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-scripts" (OuterVolumeSpecName: "scripts") pod "8d699fdc-91ac-446c-812c-2dd9f5045947" (UID: "8d699fdc-91ac-446c-812c-2dd9f5045947"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.116556 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-config-data" (OuterVolumeSpecName: "config-data") pod "8d699fdc-91ac-446c-812c-2dd9f5045947" (UID: "8d699fdc-91ac-446c-812c-2dd9f5045947"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.117571 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d699fdc-91ac-446c-812c-2dd9f5045947" (UID: "8d699fdc-91ac-446c-812c-2dd9f5045947"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.185479 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.185558 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwdjx\" (UniqueName: \"kubernetes.io/projected/8d699fdc-91ac-446c-812c-2dd9f5045947-kube-api-access-vwdjx\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.185578 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.185615 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.185631 4706 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.185643 4706 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d699fdc-91ac-446c-812c-2dd9f5045947-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.391241 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rrxz6" Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.391463 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rrxz6" event={"ID":"8d699fdc-91ac-446c-812c-2dd9f5045947","Type":"ContainerDied","Data":"48ae9e94636771901d87d2296b2e7cdf9393df96a3446da6c221e8d851ce33e2"} Dec 08 19:42:27 crc kubenswrapper[4706]: I1208 19:42:27.391810 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48ae9e94636771901d87d2296b2e7cdf9393df96a3446da6c221e8d851ce33e2" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.160766 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rrxz6"] Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.172138 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rrxz6"] Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.261924 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jhpwq"] Dec 08 19:42:28 crc kubenswrapper[4706]: E1208 19:42:28.262610 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d699fdc-91ac-446c-812c-2dd9f5045947" containerName="keystone-bootstrap" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.262627 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d699fdc-91ac-446c-812c-2dd9f5045947" containerName="keystone-bootstrap" Dec 08 19:42:28 crc kubenswrapper[4706]: E1208 19:42:28.262638 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e850531-c236-429d-abdc-ab69df9a82a5" containerName="dnsmasq-dns" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.262644 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e850531-c236-429d-abdc-ab69df9a82a5" containerName="dnsmasq-dns" Dec 08 19:42:28 crc kubenswrapper[4706]: E1208 19:42:28.263869 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e850531-c236-429d-abdc-ab69df9a82a5" containerName="init" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.263888 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e850531-c236-429d-abdc-ab69df9a82a5" containerName="init" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.264196 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d699fdc-91ac-446c-812c-2dd9f5045947" containerName="keystone-bootstrap" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.264226 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e850531-c236-429d-abdc-ab69df9a82a5" containerName="dnsmasq-dns" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.265137 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.269548 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.269696 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.269885 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wbnw9" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.271094 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.279005 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jhpwq"] Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.312834 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-credential-keys\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.313006 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-config-data\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.313065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mljb5\" (UniqueName: \"kubernetes.io/projected/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-kube-api-access-mljb5\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.313109 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-scripts\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.313190 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-combined-ca-bundle\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.313218 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-fernet-keys\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.415157 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-config-data\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.415292 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mljb5\" (UniqueName: \"kubernetes.io/projected/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-kube-api-access-mljb5\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.415355 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-scripts\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.415434 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-combined-ca-bundle\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.415462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-fernet-keys\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.415502 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-credential-keys\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.424211 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-config-data\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.436130 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-scripts\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.436310 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-credential-keys\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.439019 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-combined-ca-bundle\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.440670 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-fernet-keys\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.440880 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mljb5\" (UniqueName: \"kubernetes.io/projected/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-kube-api-access-mljb5\") pod \"keystone-bootstrap-jhpwq\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: I1208 19:42:28.596766 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:28 crc kubenswrapper[4706]: E1208 19:42:28.968711 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 08 19:42:28 crc kubenswrapper[4706]: E1208 19:42:28.969077 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t6s9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-brz4k_openstack(776cb4c2-52c5-404a-b3ca-c83f93dcbd0e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:42:28 crc kubenswrapper[4706]: E1208 19:42:28.970334 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-brz4k" podUID="776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" Dec 08 19:42:29 crc kubenswrapper[4706]: E1208 19:42:29.417975 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-brz4k" podUID="776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" Dec 08 19:42:29 crc kubenswrapper[4706]: I1208 19:42:29.625633 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d699fdc-91ac-446c-812c-2dd9f5045947" path="/var/lib/kubelet/pods/8d699fdc-91ac-446c-812c-2dd9f5045947/volumes" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.433729 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.444049 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dkmkk" event={"ID":"18ef7995-14d2-4c7a-8b30-8aaac1c49b92","Type":"ContainerDied","Data":"100b5d19b5b82eb64d96177a54638bdae0dae7b127709a9d3c7a7934cdfbb76b"} Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.456432 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="100b5d19b5b82eb64d96177a54638bdae0dae7b127709a9d3c7a7934cdfbb76b" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.456581 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dkmkk" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.499195 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj2f5\" (UniqueName: \"kubernetes.io/projected/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-kube-api-access-sj2f5\") pod \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.500249 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-config\") pod \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.500352 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-combined-ca-bundle\") pod \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\" (UID: \"18ef7995-14d2-4c7a-8b30-8aaac1c49b92\") " Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.510560 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-kube-api-access-sj2f5" (OuterVolumeSpecName: "kube-api-access-sj2f5") pod "18ef7995-14d2-4c7a-8b30-8aaac1c49b92" (UID: "18ef7995-14d2-4c7a-8b30-8aaac1c49b92"). InnerVolumeSpecName "kube-api-access-sj2f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.564582 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18ef7995-14d2-4c7a-8b30-8aaac1c49b92" (UID: "18ef7995-14d2-4c7a-8b30-8aaac1c49b92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.569526 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-config" (OuterVolumeSpecName: "config") pod "18ef7995-14d2-4c7a-8b30-8aaac1c49b92" (UID: "18ef7995-14d2-4c7a-8b30-8aaac1c49b92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.604125 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj2f5\" (UniqueName: \"kubernetes.io/projected/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-kube-api-access-sj2f5\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.604167 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:31 crc kubenswrapper[4706]: I1208 19:42:31.604182 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ef7995-14d2-4c7a-8b30-8aaac1c49b92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.470329 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" event={"ID":"e87e3ba7-b301-4df4-9ca5-2f3b735c622f","Type":"ContainerStarted","Data":"72b4e59a6fcad9c7749dc6044b0523bc4319ff11361e399682b44cd1f020672c"} Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.471916 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.516942 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" podStartSLOduration=31.516910352 podStartE2EDuration="31.516910352s" podCreationTimestamp="2025-12-08 19:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:32.494888691 +0000 UTC m=+1235.137089694" watchObservedRunningTime="2025-12-08 19:42:32.516910352 +0000 UTC m=+1235.159111355" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.647714 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c86wj"] Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.741836 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d88d7b95f-82ttp"] Dec 08 19:42:32 crc kubenswrapper[4706]: E1208 19:42:32.742492 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ef7995-14d2-4c7a-8b30-8aaac1c49b92" containerName="neutron-db-sync" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.742517 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ef7995-14d2-4c7a-8b30-8aaac1c49b92" containerName="neutron-db-sync" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.742890 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ef7995-14d2-4c7a-8b30-8aaac1c49b92" containerName="neutron-db-sync" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.744472 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.779321 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d88d7b95f-82ttp"] Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.800375 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7bbb444974-g2ncm"] Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.803705 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.806399 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.806676 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.811496 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8p7hk" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.820103 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.837251 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-ovndb-tls-certs\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.837891 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-combined-ca-bundle\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.838028 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-config\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.838158 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-httpd-config\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.838442 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-swift-storage-0\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.838545 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-nb\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.838701 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2gqz\" (UniqueName: \"kubernetes.io/projected/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-kube-api-access-g2gqz\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.838803 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-sb\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.838932 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-config\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.839335 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-svc\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.843686 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh7qm\" (UniqueName: \"kubernetes.io/projected/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-kube-api-access-bh7qm\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.854840 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bbb444974-g2ncm"] Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.946080 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-swift-storage-0\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947073 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-nb\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947135 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2gqz\" (UniqueName: \"kubernetes.io/projected/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-kube-api-access-g2gqz\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947165 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-sb\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947190 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-config\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947273 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-svc\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947302 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh7qm\" (UniqueName: \"kubernetes.io/projected/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-kube-api-access-bh7qm\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947342 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-ovndb-tls-certs\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947382 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-combined-ca-bundle\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947414 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-config\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947437 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-httpd-config\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.947017 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-swift-storage-0\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.948941 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-svc\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.949381 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-config\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.949431 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-sb\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.952019 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-nb\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.954204 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-config\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.954922 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-httpd-config\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.955934 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-ovndb-tls-certs\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.956986 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-combined-ca-bundle\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.967878 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2gqz\" (UniqueName: \"kubernetes.io/projected/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-kube-api-access-g2gqz\") pod \"neutron-7bbb444974-g2ncm\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:32 crc kubenswrapper[4706]: I1208 19:42:32.976152 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh7qm\" (UniqueName: \"kubernetes.io/projected/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-kube-api-access-bh7qm\") pod \"dnsmasq-dns-7d88d7b95f-82ttp\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:33 crc kubenswrapper[4706]: I1208 19:42:33.079727 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:33 crc kubenswrapper[4706]: I1208 19:42:33.131986 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:34 crc kubenswrapper[4706]: I1208 19:42:34.488709 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" podUID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" containerName="dnsmasq-dns" containerID="cri-o://72b4e59a6fcad9c7749dc6044b0523bc4319ff11361e399682b44cd1f020672c" gracePeriod=10 Dec 08 19:42:35 crc kubenswrapper[4706]: E1208 19:42:35.187228 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Dec 08 19:42:35 crc kubenswrapper[4706]: E1208 19:42:35.187337 4706 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Dec 08 19:42:35 crc kubenswrapper[4706]: E1208 19:42:35.187550 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-78d8k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42406,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-59fls_openstack(e68efceb-8175-4600-bc86-0d9d5d9d5dbd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:42:35 crc kubenswrapper[4706]: E1208 19:42:35.189067 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-59fls" podUID="e68efceb-8175-4600-bc86-0d9d5d9d5dbd" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.263776 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-79bc9d977-gw6pk"] Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.272070 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.286842 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.287137 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.317890 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-ovndb-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.318054 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-public-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.318737 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-config\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.319001 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-combined-ca-bundle\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.319062 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-internal-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.319163 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-httpd-config\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.319348 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgc7b\" (UniqueName: \"kubernetes.io/projected/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-kube-api-access-fgc7b\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.323652 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79bc9d977-gw6pk"] Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.423877 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-config\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.423956 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-combined-ca-bundle\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.423984 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-internal-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.424019 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-httpd-config\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.424063 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgc7b\" (UniqueName: \"kubernetes.io/projected/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-kube-api-access-fgc7b\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.424156 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-ovndb-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.424186 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-public-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.435214 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-ovndb-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.437559 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-internal-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.440753 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-httpd-config\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.440890 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-config\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.440808 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-public-tls-certs\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.444711 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgc7b\" (UniqueName: \"kubernetes.io/projected/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-kube-api-access-fgc7b\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.452936 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43-combined-ca-bundle\") pod \"neutron-79bc9d977-gw6pk\" (UID: \"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43\") " pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.540072 4706 generic.go:334] "Generic (PLEG): container finished" podID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" containerID="72b4e59a6fcad9c7749dc6044b0523bc4319ff11361e399682b44cd1f020672c" exitCode=0 Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.540234 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" event={"ID":"e87e3ba7-b301-4df4-9ca5-2f3b735c622f","Type":"ContainerDied","Data":"72b4e59a6fcad9c7749dc6044b0523bc4319ff11361e399682b44cd1f020672c"} Dec 08 19:42:35 crc kubenswrapper[4706]: E1208 19:42:35.546239 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current\\\"\"" pod="openstack/cloudkitty-db-sync-59fls" podUID="e68efceb-8175-4600-bc86-0d9d5d9d5dbd" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.612103 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.835814 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.835891 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:42:35 crc kubenswrapper[4706]: I1208 19:42:35.851291 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jhpwq"] Dec 08 19:42:36 crc kubenswrapper[4706]: W1208 19:42:36.033713 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2dc07a4_5f58_43b2_918c_ea0a5fbb2b6d.slice/crio-4b13c3827cbd380781b3e67c2324b06d1b77d9df1506124951d4cf6248b2c69e WatchSource:0}: Error finding container 4b13c3827cbd380781b3e67c2324b06d1b77d9df1506124951d4cf6248b2c69e: Status 404 returned error can't find the container with id 4b13c3827cbd380781b3e67c2324b06d1b77d9df1506124951d4cf6248b2c69e Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.121517 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.142679 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-sb\") pod \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.142777 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-nb\") pod \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.142986 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-swift-storage-0\") pod \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.143174 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrxtw\" (UniqueName: \"kubernetes.io/projected/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-kube-api-access-qrxtw\") pod \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.143280 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-config\") pod \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.143353 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-svc\") pod \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\" (UID: \"e87e3ba7-b301-4df4-9ca5-2f3b735c622f\") " Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.161327 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-kube-api-access-qrxtw" (OuterVolumeSpecName: "kube-api-access-qrxtw") pod "e87e3ba7-b301-4df4-9ca5-2f3b735c622f" (UID: "e87e3ba7-b301-4df4-9ca5-2f3b735c622f"). InnerVolumeSpecName "kube-api-access-qrxtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.246010 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrxtw\" (UniqueName: \"kubernetes.io/projected/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-kube-api-access-qrxtw\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.256932 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-config" (OuterVolumeSpecName: "config") pod "e87e3ba7-b301-4df4-9ca5-2f3b735c622f" (UID: "e87e3ba7-b301-4df4-9ca5-2f3b735c622f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.297633 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e87e3ba7-b301-4df4-9ca5-2f3b735c622f" (UID: "e87e3ba7-b301-4df4-9ca5-2f3b735c622f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.311707 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e87e3ba7-b301-4df4-9ca5-2f3b735c622f" (UID: "e87e3ba7-b301-4df4-9ca5-2f3b735c622f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.349333 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.349374 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.349388 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.360143 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e87e3ba7-b301-4df4-9ca5-2f3b735c622f" (UID: "e87e3ba7-b301-4df4-9ca5-2f3b735c622f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.362849 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e87e3ba7-b301-4df4-9ca5-2f3b735c622f" (UID: "e87e3ba7-b301-4df4-9ca5-2f3b735c622f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.451438 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.451476 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e87e3ba7-b301-4df4-9ca5-2f3b735c622f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.647795 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" event={"ID":"e87e3ba7-b301-4df4-9ca5-2f3b735c622f","Type":"ContainerDied","Data":"19a49a12654ad4d7464703cd6fbce9deff82c446454e90d89a6cef77fb367b0f"} Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.647871 4706 scope.go:117] "RemoveContainer" containerID="72b4e59a6fcad9c7749dc6044b0523bc4319ff11361e399682b44cd1f020672c" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.648051 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c86wj" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.707784 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jhpwq" event={"ID":"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d","Type":"ContainerStarted","Data":"4b13c3827cbd380781b3e67c2324b06d1b77d9df1506124951d4cf6248b2c69e"} Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.857476 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d88d7b95f-82ttp"] Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.880965 4706 scope.go:117] "RemoveContainer" containerID="ddd93e30d02810dea6a7c9829ce9248c26bdd714b7214f2dcd3cefd454732f81" Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.892234 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c86wj"] Dec 08 19:42:36 crc kubenswrapper[4706]: I1208 19:42:36.917367 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c86wj"] Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.177413 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.256058 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bbb444974-g2ncm"] Dec 08 19:42:37 crc kubenswrapper[4706]: W1208 19:42:37.276006 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05f612d9_fc1c_4dcd_9ccf_9eb7c41fd147.slice/crio-b732a5ae92086a1ffe077f23d24d56616cf2296b07c236ded6a8b9b2ae46b7d6 WatchSource:0}: Error finding container b732a5ae92086a1ffe077f23d24d56616cf2296b07c236ded6a8b9b2ae46b7d6: Status 404 returned error can't find the container with id b732a5ae92086a1ffe077f23d24d56616cf2296b07c236ded6a8b9b2ae46b7d6 Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.627252 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" path="/var/lib/kubelet/pods/e87e3ba7-b301-4df4-9ca5-2f3b735c622f/volumes" Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.820542 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79bc9d977-gw6pk"] Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.822123 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a298368b-7b89-4c7a-8427-330e9d16ea00","Type":"ContainerStarted","Data":"6b59caf3b749aaf5fe740648eb1227d7639d865de8dba1b4f12c5c386f63dd51"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.830696 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ff9kc" event={"ID":"8677c722-159f-4509-a1b1-114817ef241b","Type":"ContainerStarted","Data":"c82b5de45969cfe9bdd14bd26c0cf70ebdc1ca0cf543a3250b7973a433f31d1c"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.839724 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbb444974-g2ncm" event={"ID":"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147","Type":"ContainerStarted","Data":"5e3995ca1f1359ced582de216ad8ebc70c7f73aed776be390e16bce54a023321"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.839786 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbb444974-g2ncm" event={"ID":"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147","Type":"ContainerStarted","Data":"b732a5ae92086a1ffe077f23d24d56616cf2296b07c236ded6a8b9b2ae46b7d6"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.851850 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jhpwq" event={"ID":"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d","Type":"ContainerStarted","Data":"62ff1d077e6f82a9f25fc066972b37cf8127ed601d34a18acbb4ddfb98445689"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.856889 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ff9kc" podStartSLOduration=9.612033102 podStartE2EDuration="45.85686292s" podCreationTimestamp="2025-12-08 19:41:52 +0000 UTC" firstStartedPulling="2025-12-08 19:41:55.071346777 +0000 UTC m=+1197.713547780" lastFinishedPulling="2025-12-08 19:42:31.316176595 +0000 UTC m=+1233.958377598" observedRunningTime="2025-12-08 19:42:37.848556292 +0000 UTC m=+1240.490757295" watchObservedRunningTime="2025-12-08 19:42:37.85686292 +0000 UTC m=+1240.499063923" Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.859866 4706 generic.go:334] "Generic (PLEG): container finished" podID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerID="b27f249bc9b1e5e681913b0b7145807cc1ff04cde87db165059ee5be72765df2" exitCode=0 Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.859995 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" event={"ID":"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce","Type":"ContainerDied","Data":"b27f249bc9b1e5e681913b0b7145807cc1ff04cde87db165059ee5be72765df2"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.860076 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" event={"ID":"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce","Type":"ContainerStarted","Data":"050c3c79097bad4386b5fb14a9a2a41d6ee36c162afbd4d5f8c99da9809a3f36"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.870590 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8s7bd" event={"ID":"e18991e1-84d5-4e32-9a05-ad08f31cbd60","Type":"ContainerStarted","Data":"bb6f1a9e15babd2f25e5b3f17b0c5392aa16c1ca08b9fc1aeac629c0e3e9de33"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.887938 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jhpwq" podStartSLOduration=9.881800204 podStartE2EDuration="9.881800204s" podCreationTimestamp="2025-12-08 19:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:37.872960001 +0000 UTC m=+1240.515161014" watchObservedRunningTime="2025-12-08 19:42:37.881800204 +0000 UTC m=+1240.524001207" Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.895096 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lhc9c" event={"ID":"d6007166-5c77-46c7-8c2a-e4ff780725d6","Type":"ContainerStarted","Data":"c9032f02ec92f5cce032c294f93e9863342b56a15ee19e00d84c0f13dd8145e1"} Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.916850 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-8s7bd" podStartSLOduration=13.340643161 podStartE2EDuration="1m14.916822567s" podCreationTimestamp="2025-12-08 19:41:23 +0000 UTC" firstStartedPulling="2025-12-08 19:41:26.117581581 +0000 UTC m=+1168.759782584" lastFinishedPulling="2025-12-08 19:42:27.693760987 +0000 UTC m=+1230.335961990" observedRunningTime="2025-12-08 19:42:37.90017112 +0000 UTC m=+1240.542372123" watchObservedRunningTime="2025-12-08 19:42:37.916822567 +0000 UTC m=+1240.559023570" Dec 08 19:42:37 crc kubenswrapper[4706]: I1208 19:42:37.995738 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-lhc9c" podStartSLOduration=4.970510426 podStartE2EDuration="45.995709816s" podCreationTimestamp="2025-12-08 19:41:52 +0000 UTC" firstStartedPulling="2025-12-08 19:41:55.063529073 +0000 UTC m=+1197.705730076" lastFinishedPulling="2025-12-08 19:42:36.088728463 +0000 UTC m=+1238.730929466" observedRunningTime="2025-12-08 19:42:37.952395815 +0000 UTC m=+1240.594596808" watchObservedRunningTime="2025-12-08 19:42:37.995709816 +0000 UTC m=+1240.637910829" Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.917792 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" event={"ID":"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce","Type":"ContainerStarted","Data":"1be6dd51e03f7d43b1a1729d0c885697e581e860d2f114e8ec39e783fe37d36c"} Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.919123 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.922821 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bc9d977-gw6pk" event={"ID":"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43","Type":"ContainerStarted","Data":"e6baeccecfc85cf3af3e8e517a7c4de9c33bec5a452d629e436d6a90270c5a3a"} Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.922968 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bc9d977-gw6pk" event={"ID":"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43","Type":"ContainerStarted","Data":"553738f8154ea7e5a5a0c004d9d2f4f94ab263b06d11b2abd337007062a5b665"} Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.922985 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bc9d977-gw6pk" event={"ID":"94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43","Type":"ContainerStarted","Data":"c3dbee0b51dbcee787520073ad5e1d0676c27d7c84973b63ef6bd7b3430298bd"} Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.923355 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.935112 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbb444974-g2ncm" event={"ID":"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147","Type":"ContainerStarted","Data":"4a8e5955c775d933fb66a0b03ba1f45d3189af279c759b6ad3ef4b069678bc77"} Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.936406 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.952446 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" podStartSLOduration=6.952418063 podStartE2EDuration="6.952418063s" podCreationTimestamp="2025-12-08 19:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:38.940662007 +0000 UTC m=+1241.582863010" watchObservedRunningTime="2025-12-08 19:42:38.952418063 +0000 UTC m=+1241.594619066" Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.977106 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-79bc9d977-gw6pk" podStartSLOduration=3.977075749 podStartE2EDuration="3.977075749s" podCreationTimestamp="2025-12-08 19:42:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:38.965944511 +0000 UTC m=+1241.608145514" watchObservedRunningTime="2025-12-08 19:42:38.977075749 +0000 UTC m=+1241.619276772" Dec 08 19:42:38 crc kubenswrapper[4706]: I1208 19:42:38.998980 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7bbb444974-g2ncm" podStartSLOduration=6.998948346 podStartE2EDuration="6.998948346s" podCreationTimestamp="2025-12-08 19:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:38.993920492 +0000 UTC m=+1241.636121495" watchObservedRunningTime="2025-12-08 19:42:38.998948346 +0000 UTC m=+1241.641149349" Dec 08 19:42:43 crc kubenswrapper[4706]: I1208 19:42:43.082539 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:42:43 crc kubenswrapper[4706]: I1208 19:42:43.154010 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-vs9wr"] Dec 08 19:42:43 crc kubenswrapper[4706]: I1208 19:42:43.154350 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-vs9wr" podUID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerName="dnsmasq-dns" containerID="cri-o://c1b660ef7677eefaf41a2b192223ca816a3de8b32ee1f06fe79f22e5b7435ee3" gracePeriod=10 Dec 08 19:42:46 crc kubenswrapper[4706]: I1208 19:42:46.026366 4706 generic.go:334] "Generic (PLEG): container finished" podID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerID="c1b660ef7677eefaf41a2b192223ca816a3de8b32ee1f06fe79f22e5b7435ee3" exitCode=0 Dec 08 19:42:46 crc kubenswrapper[4706]: I1208 19:42:46.026467 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-vs9wr" event={"ID":"b17a1b80-f5a5-411a-a5bd-79aa80bed406","Type":"ContainerDied","Data":"c1b660ef7677eefaf41a2b192223ca816a3de8b32ee1f06fe79f22e5b7435ee3"} Dec 08 19:42:47 crc kubenswrapper[4706]: I1208 19:42:47.042047 4706 generic.go:334] "Generic (PLEG): container finished" podID="c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" containerID="62ff1d077e6f82a9f25fc066972b37cf8127ed601d34a18acbb4ddfb98445689" exitCode=0 Dec 08 19:42:47 crc kubenswrapper[4706]: I1208 19:42:47.042123 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jhpwq" event={"ID":"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d","Type":"ContainerDied","Data":"62ff1d077e6f82a9f25fc066972b37cf8127ed601d34a18acbb4ddfb98445689"} Dec 08 19:42:47 crc kubenswrapper[4706]: I1208 19:42:47.044446 4706 generic.go:334] "Generic (PLEG): container finished" podID="d6007166-5c77-46c7-8c2a-e4ff780725d6" containerID="c9032f02ec92f5cce032c294f93e9863342b56a15ee19e00d84c0f13dd8145e1" exitCode=0 Dec 08 19:42:47 crc kubenswrapper[4706]: I1208 19:42:47.044471 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lhc9c" event={"ID":"d6007166-5c77-46c7-8c2a-e4ff780725d6","Type":"ContainerDied","Data":"c9032f02ec92f5cce032c294f93e9863342b56a15ee19e00d84c0f13dd8145e1"} Dec 08 19:42:48 crc kubenswrapper[4706]: I1208 19:42:48.061933 4706 generic.go:334] "Generic (PLEG): container finished" podID="8677c722-159f-4509-a1b1-114817ef241b" containerID="c82b5de45969cfe9bdd14bd26c0cf70ebdc1ca0cf543a3250b7973a433f31d1c" exitCode=0 Dec 08 19:42:48 crc kubenswrapper[4706]: I1208 19:42:48.062136 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ff9kc" event={"ID":"8677c722-159f-4509-a1b1-114817ef241b","Type":"ContainerDied","Data":"c82b5de45969cfe9bdd14bd26c0cf70ebdc1ca0cf543a3250b7973a433f31d1c"} Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.732473 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.745498 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.760521 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lhc9c" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.775344 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.858776 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-scripts\") pod \"d6007166-5c77-46c7-8c2a-e4ff780725d6\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.858881 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-config\") pod \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.858908 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-dns-svc\") pod \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.858958 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnlsq\" (UniqueName: \"kubernetes.io/projected/b17a1b80-f5a5-411a-a5bd-79aa80bed406-kube-api-access-rnlsq\") pod \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859044 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-credential-keys\") pod \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859078 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-fernet-keys\") pod \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859115 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-combined-ca-bundle\") pod \"8677c722-159f-4509-a1b1-114817ef241b\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859142 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-combined-ca-bundle\") pod \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859184 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-config-data\") pod \"d6007166-5c77-46c7-8c2a-e4ff780725d6\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859290 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-db-sync-config-data\") pod \"8677c722-159f-4509-a1b1-114817ef241b\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859314 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-sb\") pod \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859338 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-nb\") pod \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\" (UID: \"b17a1b80-f5a5-411a-a5bd-79aa80bed406\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859357 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm24w\" (UniqueName: \"kubernetes.io/projected/d6007166-5c77-46c7-8c2a-e4ff780725d6-kube-api-access-sm24w\") pod \"d6007166-5c77-46c7-8c2a-e4ff780725d6\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859436 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-config-data\") pod \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859472 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkwnh\" (UniqueName: \"kubernetes.io/projected/8677c722-159f-4509-a1b1-114817ef241b-kube-api-access-fkwnh\") pod \"8677c722-159f-4509-a1b1-114817ef241b\" (UID: \"8677c722-159f-4509-a1b1-114817ef241b\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859522 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6007166-5c77-46c7-8c2a-e4ff780725d6-logs\") pod \"d6007166-5c77-46c7-8c2a-e4ff780725d6\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859555 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mljb5\" (UniqueName: \"kubernetes.io/projected/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-kube-api-access-mljb5\") pod \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859676 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-combined-ca-bundle\") pod \"d6007166-5c77-46c7-8c2a-e4ff780725d6\" (UID: \"d6007166-5c77-46c7-8c2a-e4ff780725d6\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.859695 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-scripts\") pod \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\" (UID: \"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d\") " Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.870640 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-scripts" (OuterVolumeSpecName: "scripts") pod "d6007166-5c77-46c7-8c2a-e4ff780725d6" (UID: "d6007166-5c77-46c7-8c2a-e4ff780725d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.872004 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b17a1b80-f5a5-411a-a5bd-79aa80bed406-kube-api-access-rnlsq" (OuterVolumeSpecName: "kube-api-access-rnlsq") pod "b17a1b80-f5a5-411a-a5bd-79aa80bed406" (UID: "b17a1b80-f5a5-411a-a5bd-79aa80bed406"). InnerVolumeSpecName "kube-api-access-rnlsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.872179 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6007166-5c77-46c7-8c2a-e4ff780725d6-logs" (OuterVolumeSpecName: "logs") pod "d6007166-5c77-46c7-8c2a-e4ff780725d6" (UID: "d6007166-5c77-46c7-8c2a-e4ff780725d6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.872462 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8677c722-159f-4509-a1b1-114817ef241b-kube-api-access-fkwnh" (OuterVolumeSpecName: "kube-api-access-fkwnh") pod "8677c722-159f-4509-a1b1-114817ef241b" (UID: "8677c722-159f-4509-a1b1-114817ef241b"). InnerVolumeSpecName "kube-api-access-fkwnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.872468 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-scripts" (OuterVolumeSpecName: "scripts") pod "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" (UID: "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.873862 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" (UID: "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.877112 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" (UID: "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.879202 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6007166-5c77-46c7-8c2a-e4ff780725d6-kube-api-access-sm24w" (OuterVolumeSpecName: "kube-api-access-sm24w") pod "d6007166-5c77-46c7-8c2a-e4ff780725d6" (UID: "d6007166-5c77-46c7-8c2a-e4ff780725d6"). InnerVolumeSpecName "kube-api-access-sm24w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.879959 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-kube-api-access-mljb5" (OuterVolumeSpecName: "kube-api-access-mljb5") pod "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" (UID: "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d"). InnerVolumeSpecName "kube-api-access-mljb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.881274 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8677c722-159f-4509-a1b1-114817ef241b" (UID: "8677c722-159f-4509-a1b1-114817ef241b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.925526 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6007166-5c77-46c7-8c2a-e4ff780725d6" (UID: "d6007166-5c77-46c7-8c2a-e4ff780725d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.944306 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b17a1b80-f5a5-411a-a5bd-79aa80bed406" (UID: "b17a1b80-f5a5-411a-a5bd-79aa80bed406"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.949793 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-config-data" (OuterVolumeSpecName: "config-data") pod "d6007166-5c77-46c7-8c2a-e4ff780725d6" (UID: "d6007166-5c77-46c7-8c2a-e4ff780725d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.954053 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8677c722-159f-4509-a1b1-114817ef241b" (UID: "8677c722-159f-4509-a1b1-114817ef241b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.967805 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" (UID: "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971486 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971528 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971586 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971599 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971609 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnlsq\" (UniqueName: \"kubernetes.io/projected/b17a1b80-f5a5-411a-a5bd-79aa80bed406-kube-api-access-rnlsq\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971623 4706 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971731 4706 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971744 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971755 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971764 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6007166-5c77-46c7-8c2a-e4ff780725d6-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971772 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8677c722-159f-4509-a1b1-114817ef241b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.971943 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm24w\" (UniqueName: \"kubernetes.io/projected/d6007166-5c77-46c7-8c2a-e4ff780725d6-kube-api-access-sm24w\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.980652 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-config" (OuterVolumeSpecName: "config") pod "b17a1b80-f5a5-411a-a5bd-79aa80bed406" (UID: "b17a1b80-f5a5-411a-a5bd-79aa80bed406"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.983054 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b17a1b80-f5a5-411a-a5bd-79aa80bed406" (UID: "b17a1b80-f5a5-411a-a5bd-79aa80bed406"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.993749 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkwnh\" (UniqueName: \"kubernetes.io/projected/8677c722-159f-4509-a1b1-114817ef241b-kube-api-access-fkwnh\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.994392 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-config-data" (OuterVolumeSpecName: "config-data") pod "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" (UID: "c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.997446 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6007166-5c77-46c7-8c2a-e4ff780725d6-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:49 crc kubenswrapper[4706]: I1208 19:42:49.999342 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mljb5\" (UniqueName: \"kubernetes.io/projected/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-kube-api-access-mljb5\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.040047 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b17a1b80-f5a5-411a-a5bd-79aa80bed406" (UID: "b17a1b80-f5a5-411a-a5bd-79aa80bed406"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.082401 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lhc9c" event={"ID":"d6007166-5c77-46c7-8c2a-e4ff780725d6","Type":"ContainerDied","Data":"8a86b8a328cdf3e06ff643f32cd2376d55387dc4b32f73ed547ac12d4dafd781"} Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.082449 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a86b8a328cdf3e06ff643f32cd2376d55387dc4b32f73ed547ac12d4dafd781" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.082537 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lhc9c" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.095891 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ff9kc" event={"ID":"8677c722-159f-4509-a1b1-114817ef241b","Type":"ContainerDied","Data":"b371a3f304c0de07cca76b5b6f742fd0f48623c7a76df0a4bc4428e452067606"} Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.096081 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b371a3f304c0de07cca76b5b6f742fd0f48623c7a76df0a4bc4428e452067606" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.096345 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ff9kc" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.115319 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.115388 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.115405 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.115419 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17a1b80-f5a5-411a-a5bd-79aa80bed406-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.129255 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jhpwq" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.129235 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jhpwq" event={"ID":"c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d","Type":"ContainerDied","Data":"4b13c3827cbd380781b3e67c2324b06d1b77d9df1506124951d4cf6248b2c69e"} Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.129499 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b13c3827cbd380781b3e67c2324b06d1b77d9df1506124951d4cf6248b2c69e" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.138523 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-vs9wr" event={"ID":"b17a1b80-f5a5-411a-a5bd-79aa80bed406","Type":"ContainerDied","Data":"bb5bd212bf1b0aa5921edac41b16fb83f0c136cfec36118fa768077684cd2081"} Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.138592 4706 scope.go:117] "RemoveContainer" containerID="c1b660ef7677eefaf41a2b192223ca816a3de8b32ee1f06fe79f22e5b7435ee3" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.138849 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-vs9wr" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.195875 4706 scope.go:117] "RemoveContainer" containerID="a8533ddb4be48988ddd60b6dd0daa9ea77760010e7a4f004a6979806bb7e0ec6" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.218200 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-vs9wr"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.229326 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-vs9wr"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.405510 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5b989c8b75-k95kd"] Dec 08 19:42:50 crc kubenswrapper[4706]: E1208 19:42:50.406466 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerName="dnsmasq-dns" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.406492 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerName="dnsmasq-dns" Dec 08 19:42:50 crc kubenswrapper[4706]: E1208 19:42:50.406510 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerName="init" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.406517 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerName="init" Dec 08 19:42:50 crc kubenswrapper[4706]: E1208 19:42:50.406528 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8677c722-159f-4509-a1b1-114817ef241b" containerName="barbican-db-sync" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.406536 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8677c722-159f-4509-a1b1-114817ef241b" containerName="barbican-db-sync" Dec 08 19:42:50 crc kubenswrapper[4706]: E1208 19:42:50.406545 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" containerName="dnsmasq-dns" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.406554 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" containerName="dnsmasq-dns" Dec 08 19:42:50 crc kubenswrapper[4706]: E1208 19:42:50.406578 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" containerName="init" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.406586 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" containerName="init" Dec 08 19:42:50 crc kubenswrapper[4706]: E1208 19:42:50.406596 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" containerName="keystone-bootstrap" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.406606 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" containerName="keystone-bootstrap" Dec 08 19:42:50 crc kubenswrapper[4706]: E1208 19:42:50.406637 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6007166-5c77-46c7-8c2a-e4ff780725d6" containerName="placement-db-sync" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.406645 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6007166-5c77-46c7-8c2a-e4ff780725d6" containerName="placement-db-sync" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.408850 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e87e3ba7-b301-4df4-9ca5-2f3b735c622f" containerName="dnsmasq-dns" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.408897 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8677c722-159f-4509-a1b1-114817ef241b" containerName="barbican-db-sync" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.408913 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6007166-5c77-46c7-8c2a-e4ff780725d6" containerName="placement-db-sync" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.408951 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerName="dnsmasq-dns" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.408966 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" containerName="keystone-bootstrap" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.415492 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.428304 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-c4h7r" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.428772 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.428996 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.434782 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b989c8b75-k95kd"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.536162 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-788468f6bb-gm7rb"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.537900 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-config-data-custom\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.538010 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-combined-ca-bundle\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.538239 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-config-data\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.538650 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srqm8\" (UniqueName: \"kubernetes.io/projected/2cc4f4da-55de-4f4e-9e3e-099e148987a2-kube-api-access-srqm8\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.538827 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cc4f4da-55de-4f4e-9e3e-099e148987a2-logs\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.539504 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.548664 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.575401 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-788468f6bb-gm7rb"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.599535 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ff8449c8c-q6hd4"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.601635 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.622865 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-vs9wr" podUID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.635656 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff8449c8c-q6hd4"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651308 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cc4f4da-55de-4f4e-9e3e-099e148987a2-logs\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651369 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-config-data-custom\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651397 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-combined-ca-bundle\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651445 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-svc\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651485 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-config-data\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651527 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd878\" (UniqueName: \"kubernetes.io/projected/aa3a5522-ce55-49b7-a575-3f83c373b2ac-kube-api-access-bd878\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651572 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-config-data\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651593 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa3a5522-ce55-49b7-a575-3f83c373b2ac-logs\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.651620 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-combined-ca-bundle\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.652034 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cc4f4da-55de-4f4e-9e3e-099e148987a2-logs\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.654877 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-config-data-custom\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.655135 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srqm8\" (UniqueName: \"kubernetes.io/projected/2cc4f4da-55de-4f4e-9e3e-099e148987a2-kube-api-access-srqm8\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.655213 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.655277 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-swift-storage-0\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.655318 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.655347 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq56b\" (UniqueName: \"kubernetes.io/projected/67e65750-2f38-473b-9f5e-ea2edff83e6d-kube-api-access-vq56b\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.655387 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-config\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.661367 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-combined-ca-bundle\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.673140 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-config-data\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.689795 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2cc4f4da-55de-4f4e-9e3e-099e148987a2-config-data-custom\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.699724 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srqm8\" (UniqueName: \"kubernetes.io/projected/2cc4f4da-55de-4f4e-9e3e-099e148987a2-kube-api-access-srqm8\") pod \"barbican-worker-5b989c8b75-k95kd\" (UID: \"2cc4f4da-55de-4f4e-9e3e-099e148987a2\") " pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.724627 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5b87bf4766-2rxxp"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.726952 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.736003 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.758984 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5b87bf4766-2rxxp"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.760587 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-config-data-custom\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761047 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37655fed-e807-4ad5-ada9-d2c4b0204e1d-logs\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761114 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761142 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-swift-storage-0\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761177 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761205 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq56b\" (UniqueName: \"kubernetes.io/projected/67e65750-2f38-473b-9f5e-ea2edff83e6d-kube-api-access-vq56b\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761231 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-config\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761256 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data-custom\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761332 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckhr6\" (UniqueName: \"kubernetes.io/projected/37655fed-e807-4ad5-ada9-d2c4b0204e1d-kube-api-access-ckhr6\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761382 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761429 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-svc\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761483 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd878\" (UniqueName: \"kubernetes.io/projected/aa3a5522-ce55-49b7-a575-3f83c373b2ac-kube-api-access-bd878\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761534 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-config-data\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761558 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa3a5522-ce55-49b7-a575-3f83c373b2ac-logs\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761594 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-combined-ca-bundle\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.761633 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-combined-ca-bundle\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.765731 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.765851 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.766761 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-config\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.767541 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa3a5522-ce55-49b7-a575-3f83c373b2ac-logs\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.767729 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-svc\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.770441 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-config-data-custom\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.773794 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-swift-storage-0\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.777984 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b989c8b75-k95kd" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.778212 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-config-data\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.784913 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3a5522-ce55-49b7-a575-3f83c373b2ac-combined-ca-bundle\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.811831 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq56b\" (UniqueName: \"kubernetes.io/projected/67e65750-2f38-473b-9f5e-ea2edff83e6d-kube-api-access-vq56b\") pod \"dnsmasq-dns-5ff8449c8c-q6hd4\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.830463 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd878\" (UniqueName: \"kubernetes.io/projected/aa3a5522-ce55-49b7-a575-3f83c373b2ac-kube-api-access-bd878\") pod \"barbican-keystone-listener-788468f6bb-gm7rb\" (UID: \"aa3a5522-ce55-49b7-a575-3f83c373b2ac\") " pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.864625 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckhr6\" (UniqueName: \"kubernetes.io/projected/37655fed-e807-4ad5-ada9-d2c4b0204e1d-kube-api-access-ckhr6\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.864738 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.864940 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-combined-ca-bundle\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.865237 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37655fed-e807-4ad5-ada9-d2c4b0204e1d-logs\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.865383 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data-custom\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.866611 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37655fed-e807-4ad5-ada9-d2c4b0204e1d-logs\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.870860 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-combined-ca-bundle\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.871757 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data-custom\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.873603 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.903023 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckhr6\" (UniqueName: \"kubernetes.io/projected/37655fed-e807-4ad5-ada9-d2c4b0204e1d-kube-api-access-ckhr6\") pod \"barbican-api-5b87bf4766-2rxxp\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.926643 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.937205 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.937904 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.976338 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-786cfc44fd-zdf7t"] Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.978476 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.984008 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.984342 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.984540 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.987744 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wbnw9" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.987966 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.988380 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 08 19:42:50 crc kubenswrapper[4706]: I1208 19:42:50.993776 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-786cfc44fd-zdf7t"] Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.034056 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7494d94d4d-hn8rd"] Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.036167 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.048676 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bcpc2" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.048892 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.049122 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.049284 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.050519 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.062308 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7494d94d4d-hn8rd"] Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075546 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-internal-tls-certs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075601 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-scripts\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075635 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-internal-tls-certs\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075659 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-public-tls-certs\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075700 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-config-data\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075744 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgx2g\" (UniqueName: \"kubernetes.io/projected/4abc8b62-e549-41c9-8bc9-8d35f3074e79-kube-api-access-tgx2g\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075794 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-combined-ca-bundle\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075844 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-public-tls-certs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075907 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-config-data\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075932 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-combined-ca-bundle\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075965 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf4bd\" (UniqueName: \"kubernetes.io/projected/93159ff6-eda3-42c1-94db-46a5d0777a47-kube-api-access-kf4bd\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.075983 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-scripts\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.076007 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-fernet-keys\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.076026 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-credential-keys\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.076042 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4abc8b62-e549-41c9-8bc9-8d35f3074e79-logs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196204 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf4bd\" (UniqueName: \"kubernetes.io/projected/93159ff6-eda3-42c1-94db-46a5d0777a47-kube-api-access-kf4bd\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196732 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-scripts\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196771 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-credential-keys\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196800 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-fernet-keys\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196830 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4abc8b62-e549-41c9-8bc9-8d35f3074e79-logs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196918 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-internal-tls-certs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196948 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-scripts\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196975 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-internal-tls-certs\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.196996 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-public-tls-certs\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.197049 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-config-data\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.197098 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgx2g\" (UniqueName: \"kubernetes.io/projected/4abc8b62-e549-41c9-8bc9-8d35f3074e79-kube-api-access-tgx2g\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.197175 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-combined-ca-bundle\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.197195 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-public-tls-certs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.197227 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-config-data\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.197287 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-combined-ca-bundle\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.208758 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4abc8b62-e549-41c9-8bc9-8d35f3074e79-logs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.212203 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-internal-tls-certs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.222278 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-config-data\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.223086 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-public-tls-certs\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.223877 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-scripts\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.223967 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-credential-keys\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.224161 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abc8b62-e549-41c9-8bc9-8d35f3074e79-combined-ca-bundle\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.224779 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-fernet-keys\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.229655 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-scripts\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.230252 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-public-tls-certs\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.233738 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-config-data\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.233885 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf4bd\" (UniqueName: \"kubernetes.io/projected/93159ff6-eda3-42c1-94db-46a5d0777a47-kube-api-access-kf4bd\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.234804 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-internal-tls-certs\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.240993 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a298368b-7b89-4c7a-8427-330e9d16ea00","Type":"ContainerStarted","Data":"a4a69a2c1c7767822f0dea5bbb310b47cc97d60c9ab969d5765ed5bb27f7edcc"} Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.244028 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgx2g\" (UniqueName: \"kubernetes.io/projected/4abc8b62-e549-41c9-8bc9-8d35f3074e79-kube-api-access-tgx2g\") pod \"placement-7494d94d4d-hn8rd\" (UID: \"4abc8b62-e549-41c9-8bc9-8d35f3074e79\") " pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.247351 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93159ff6-eda3-42c1-94db-46a5d0777a47-combined-ca-bundle\") pod \"keystone-786cfc44fd-zdf7t\" (UID: \"93159ff6-eda3-42c1-94db-46a5d0777a47\") " pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.259754 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-59fls" event={"ID":"e68efceb-8175-4600-bc86-0d9d5d9d5dbd","Type":"ContainerStarted","Data":"891fc2499228fc87fb1799688a10e66c513f244ea0bbbe204a8c839d63f4e8ce"} Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.327969 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-59fls" podStartSLOduration=4.42984492 podStartE2EDuration="59.327939983s" podCreationTimestamp="2025-12-08 19:41:52 +0000 UTC" firstStartedPulling="2025-12-08 19:41:55.094271492 +0000 UTC m=+1197.736472495" lastFinishedPulling="2025-12-08 19:42:49.992366565 +0000 UTC m=+1252.634567558" observedRunningTime="2025-12-08 19:42:51.298692425 +0000 UTC m=+1253.940893438" watchObservedRunningTime="2025-12-08 19:42:51.327939983 +0000 UTC m=+1253.970140986" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.378398 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.381607 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.720599 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b17a1b80-f5a5-411a-a5bd-79aa80bed406" path="/var/lib/kubelet/pods/b17a1b80-f5a5-411a-a5bd-79aa80bed406/volumes" Dec 08 19:42:51 crc kubenswrapper[4706]: I1208 19:42:51.724164 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b989c8b75-k95kd"] Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.221737 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff8449c8c-q6hd4"] Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.300692 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" event={"ID":"67e65750-2f38-473b-9f5e-ea2edff83e6d","Type":"ContainerStarted","Data":"6cdcccd964f61889725c35a419e387711343358ffcd1cf6be14d607a07a852e5"} Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.310586 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-brz4k" event={"ID":"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e","Type":"ContainerStarted","Data":"843d7a7fa9e2d4430c3baedec98c6e57411f55a0075a2d7e5799758eaadae500"} Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.319440 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b989c8b75-k95kd" event={"ID":"2cc4f4da-55de-4f4e-9e3e-099e148987a2","Type":"ContainerStarted","Data":"659c46c3ac87e70a05653bcb2f5794d6b0a4f89b0e78d41fed5211775a042a3d"} Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.345501 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-brz4k" podStartSLOduration=5.667920727 podStartE2EDuration="1m0.345464833s" podCreationTimestamp="2025-12-08 19:41:52 +0000 UTC" firstStartedPulling="2025-12-08 19:41:55.239198393 +0000 UTC m=+1197.881399396" lastFinishedPulling="2025-12-08 19:42:49.916742489 +0000 UTC m=+1252.558943502" observedRunningTime="2025-12-08 19:42:52.340647635 +0000 UTC m=+1254.982848638" watchObservedRunningTime="2025-12-08 19:42:52.345464833 +0000 UTC m=+1254.987665836" Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.449775 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5b87bf4766-2rxxp"] Dec 08 19:42:52 crc kubenswrapper[4706]: W1208 19:42:52.466517 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37655fed_e807_4ad5_ada9_d2c4b0204e1d.slice/crio-43ed7cafb068d3ff39d33d3e38c438605c0865977a4f20c12f31eb5a8dda96cd WatchSource:0}: Error finding container 43ed7cafb068d3ff39d33d3e38c438605c0865977a4f20c12f31eb5a8dda96cd: Status 404 returned error can't find the container with id 43ed7cafb068d3ff39d33d3e38c438605c0865977a4f20c12f31eb5a8dda96cd Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.482309 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-788468f6bb-gm7rb"] Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.936064 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7494d94d4d-hn8rd"] Dec 08 19:42:52 crc kubenswrapper[4706]: I1208 19:42:52.971783 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-786cfc44fd-zdf7t"] Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.348595 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b87bf4766-2rxxp" event={"ID":"37655fed-e807-4ad5-ada9-d2c4b0204e1d","Type":"ContainerStarted","Data":"2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec"} Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.349677 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.349723 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b87bf4766-2rxxp" event={"ID":"37655fed-e807-4ad5-ada9-d2c4b0204e1d","Type":"ContainerStarted","Data":"d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f"} Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.349746 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b87bf4766-2rxxp" event={"ID":"37655fed-e807-4ad5-ada9-d2c4b0204e1d","Type":"ContainerStarted","Data":"43ed7cafb068d3ff39d33d3e38c438605c0865977a4f20c12f31eb5a8dda96cd"} Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.349760 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.355725 4706 generic.go:334] "Generic (PLEG): container finished" podID="67e65750-2f38-473b-9f5e-ea2edff83e6d" containerID="641ad2cc7c9eb5bf5fbf725b847b3a6c7b280039acc30c0fd3bcdfc257c1093f" exitCode=0 Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.356726 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" event={"ID":"67e65750-2f38-473b-9f5e-ea2edff83e6d","Type":"ContainerDied","Data":"641ad2cc7c9eb5bf5fbf725b847b3a6c7b280039acc30c0fd3bcdfc257c1093f"} Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.365482 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-786cfc44fd-zdf7t" event={"ID":"93159ff6-eda3-42c1-94db-46a5d0777a47","Type":"ContainerStarted","Data":"3e93e4b081005349441ddca2a3e43b3c7e58c36706b1da729ca0926e424465ed"} Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.365708 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-786cfc44fd-zdf7t" event={"ID":"93159ff6-eda3-42c1-94db-46a5d0777a47","Type":"ContainerStarted","Data":"e0d846e0a78d70a508018db081e7c0c656cc6fddf50fc20b05139173bf59a81c"} Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.372393 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.381801 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7494d94d4d-hn8rd" event={"ID":"4abc8b62-e549-41c9-8bc9-8d35f3074e79","Type":"ContainerStarted","Data":"b6ce24e2b91ca168916a7f17f7409dcce8542446fb173d8ffdd36e5c6ed5269d"} Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.384164 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" event={"ID":"aa3a5522-ce55-49b7-a575-3f83c373b2ac","Type":"ContainerStarted","Data":"9279b38ce859bce7a6fb66bb2049829dabeffc3cc9cb8d35021fd015a8a42d3c"} Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.408468 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5b87bf4766-2rxxp" podStartSLOduration=3.408438424 podStartE2EDuration="3.408438424s" podCreationTimestamp="2025-12-08 19:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:53.380158654 +0000 UTC m=+1256.022359657" watchObservedRunningTime="2025-12-08 19:42:53.408438424 +0000 UTC m=+1256.050639427" Dec 08 19:42:53 crc kubenswrapper[4706]: I1208 19:42:53.451843 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-786cfc44fd-zdf7t" podStartSLOduration=3.451816056 podStartE2EDuration="3.451816056s" podCreationTimestamp="2025-12-08 19:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:53.433245814 +0000 UTC m=+1256.075446827" watchObservedRunningTime="2025-12-08 19:42:53.451816056 +0000 UTC m=+1256.094017059" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.435558 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7494d94d4d-hn8rd" event={"ID":"4abc8b62-e549-41c9-8bc9-8d35f3074e79","Type":"ContainerStarted","Data":"95486b7d7ea52b9afbf56f3bace3ac0aec3930a5a179b4196dc9ca80a3830556"} Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.436358 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.436373 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7494d94d4d-hn8rd" event={"ID":"4abc8b62-e549-41c9-8bc9-8d35f3074e79","Type":"ContainerStarted","Data":"436e2c8fc3b0d15e194a24659090fcd06356bea20d957e1dae9d5c9ea24d78e9"} Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.436386 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.455500 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" event={"ID":"67e65750-2f38-473b-9f5e-ea2edff83e6d","Type":"ContainerStarted","Data":"9f7a8791f2990524985fb50d843433f9892dc637664d1b9bc0242b2fedf6bfc8"} Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.456615 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.461413 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7494d94d4d-hn8rd" podStartSLOduration=4.461396789 podStartE2EDuration="4.461396789s" podCreationTimestamp="2025-12-08 19:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:54.455450768 +0000 UTC m=+1257.097651791" watchObservedRunningTime="2025-12-08 19:42:54.461396789 +0000 UTC m=+1257.103597792" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.577797 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" podStartSLOduration=4.577756901 podStartE2EDuration="4.577756901s" podCreationTimestamp="2025-12-08 19:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:54.496904256 +0000 UTC m=+1257.139105259" watchObservedRunningTime="2025-12-08 19:42:54.577756901 +0000 UTC m=+1257.219957904" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.609352 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8466877bdb-zbttb"] Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.611997 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.618489 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.622543 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.666636 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-config-data-custom\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.666736 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-combined-ca-bundle\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.666770 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-public-tls-certs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.666826 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-config-data\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.666888 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0acc09f-0610-4121-b42b-ebe7274b5c5d-logs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.666949 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-internal-tls-certs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.666981 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vphbq\" (UniqueName: \"kubernetes.io/projected/b0acc09f-0610-4121-b42b-ebe7274b5c5d-kube-api-access-vphbq\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.676850 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8466877bdb-zbttb"] Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.769561 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-config-data-custom\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.769653 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-combined-ca-bundle\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.769684 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-public-tls-certs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.769744 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-config-data\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.769828 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0acc09f-0610-4121-b42b-ebe7274b5c5d-logs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.769881 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-internal-tls-certs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.769905 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vphbq\" (UniqueName: \"kubernetes.io/projected/b0acc09f-0610-4121-b42b-ebe7274b5c5d-kube-api-access-vphbq\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.772176 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0acc09f-0610-4121-b42b-ebe7274b5c5d-logs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.781350 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-public-tls-certs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.781650 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-config-data\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.781732 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-internal-tls-certs\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.783814 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-config-data-custom\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.791956 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0acc09f-0610-4121-b42b-ebe7274b5c5d-combined-ca-bundle\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.797748 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vphbq\" (UniqueName: \"kubernetes.io/projected/b0acc09f-0610-4121-b42b-ebe7274b5c5d-kube-api-access-vphbq\") pod \"barbican-api-8466877bdb-zbttb\" (UID: \"b0acc09f-0610-4121-b42b-ebe7274b5c5d\") " pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:54 crc kubenswrapper[4706]: I1208 19:42:54.936590 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.001447 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8466877bdb-zbttb"] Dec 08 19:42:56 crc kubenswrapper[4706]: W1208 19:42:56.015908 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0acc09f_0610_4121_b42b_ebe7274b5c5d.slice/crio-fe14d3f20285843d758f933668fb8c7b0edc49f460d0b08cf127dc768f397458 WatchSource:0}: Error finding container fe14d3f20285843d758f933668fb8c7b0edc49f460d0b08cf127dc768f397458: Status 404 returned error can't find the container with id fe14d3f20285843d758f933668fb8c7b0edc49f460d0b08cf127dc768f397458 Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.501964 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b989c8b75-k95kd" event={"ID":"2cc4f4da-55de-4f4e-9e3e-099e148987a2","Type":"ContainerStarted","Data":"dfd70441587f56d6f996d18a82e627fad7c7f92a9d671cceb21ffef63cb6c119"} Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.502393 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b989c8b75-k95kd" event={"ID":"2cc4f4da-55de-4f4e-9e3e-099e148987a2","Type":"ContainerStarted","Data":"05c725bf17179aed67a12dc8bd8e62e07e65d1cd59b2c43e22a67f72de25be7d"} Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.505946 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8466877bdb-zbttb" event={"ID":"b0acc09f-0610-4121-b42b-ebe7274b5c5d","Type":"ContainerStarted","Data":"2f159cab8063c2728d10f34db438d335aaeb17aab8f51662a412f40b5a8c98fd"} Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.505982 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8466877bdb-zbttb" event={"ID":"b0acc09f-0610-4121-b42b-ebe7274b5c5d","Type":"ContainerStarted","Data":"fe14d3f20285843d758f933668fb8c7b0edc49f460d0b08cf127dc768f397458"} Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.511102 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" event={"ID":"aa3a5522-ce55-49b7-a575-3f83c373b2ac","Type":"ContainerStarted","Data":"6a8bd329dbd554d2d55464b93508cfd39e24d118e6f68b2349246870767cd54a"} Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.511158 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" event={"ID":"aa3a5522-ce55-49b7-a575-3f83c373b2ac","Type":"ContainerStarted","Data":"ca3e005bcb7727ec1a837c1dc516cffe1358b9e8a09fe6c841d6944199ee15b7"} Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.530201 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5b989c8b75-k95kd" podStartSLOduration=2.917793223 podStartE2EDuration="6.530178424s" podCreationTimestamp="2025-12-08 19:42:50 +0000 UTC" firstStartedPulling="2025-12-08 19:42:51.845936838 +0000 UTC m=+1254.488137841" lastFinishedPulling="2025-12-08 19:42:55.458322039 +0000 UTC m=+1258.100523042" observedRunningTime="2025-12-08 19:42:56.521072503 +0000 UTC m=+1259.163273506" watchObservedRunningTime="2025-12-08 19:42:56.530178424 +0000 UTC m=+1259.172379427" Dec 08 19:42:56 crc kubenswrapper[4706]: I1208 19:42:56.568091 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-788468f6bb-gm7rb" podStartSLOduration=3.583502017 podStartE2EDuration="6.568066799s" podCreationTimestamp="2025-12-08 19:42:50 +0000 UTC" firstStartedPulling="2025-12-08 19:42:52.463359929 +0000 UTC m=+1255.105560932" lastFinishedPulling="2025-12-08 19:42:55.447924721 +0000 UTC m=+1258.090125714" observedRunningTime="2025-12-08 19:42:56.550225978 +0000 UTC m=+1259.192427001" watchObservedRunningTime="2025-12-08 19:42:56.568066799 +0000 UTC m=+1259.210267802" Dec 08 19:42:57 crc kubenswrapper[4706]: I1208 19:42:57.529836 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8466877bdb-zbttb" event={"ID":"b0acc09f-0610-4121-b42b-ebe7274b5c5d","Type":"ContainerStarted","Data":"f6be8545de5b6322cef04c401d4be697c2192cf7b8621047e46e034084a3f588"} Dec 08 19:42:57 crc kubenswrapper[4706]: I1208 19:42:57.564831 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8466877bdb-zbttb" podStartSLOduration=3.564797634 podStartE2EDuration="3.564797634s" podCreationTimestamp="2025-12-08 19:42:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:42:57.553101349 +0000 UTC m=+1260.195302352" watchObservedRunningTime="2025-12-08 19:42:57.564797634 +0000 UTC m=+1260.206998637" Dec 08 19:42:58 crc kubenswrapper[4706]: I1208 19:42:58.549410 4706 generic.go:334] "Generic (PLEG): container finished" podID="e68efceb-8175-4600-bc86-0d9d5d9d5dbd" containerID="891fc2499228fc87fb1799688a10e66c513f244ea0bbbe204a8c839d63f4e8ce" exitCode=0 Dec 08 19:42:58 crc kubenswrapper[4706]: I1208 19:42:58.550454 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-59fls" event={"ID":"e68efceb-8175-4600-bc86-0d9d5d9d5dbd","Type":"ContainerDied","Data":"891fc2499228fc87fb1799688a10e66c513f244ea0bbbe204a8c839d63f4e8ce"} Dec 08 19:42:58 crc kubenswrapper[4706]: I1208 19:42:58.550907 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:58 crc kubenswrapper[4706]: I1208 19:42:58.551015 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:42:59 crc kubenswrapper[4706]: I1208 19:42:59.564578 4706 generic.go:334] "Generic (PLEG): container finished" podID="e18991e1-84d5-4e32-9a05-ad08f31cbd60" containerID="bb6f1a9e15babd2f25e5b3f17b0c5392aa16c1ca08b9fc1aeac629c0e3e9de33" exitCode=0 Dec 08 19:42:59 crc kubenswrapper[4706]: I1208 19:42:59.565303 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8s7bd" event={"ID":"e18991e1-84d5-4e32-9a05-ad08f31cbd60","Type":"ContainerDied","Data":"bb6f1a9e15babd2f25e5b3f17b0c5392aa16c1ca08b9fc1aeac629c0e3e9de33"} Dec 08 19:42:59 crc kubenswrapper[4706]: I1208 19:42:59.569344 4706 generic.go:334] "Generic (PLEG): container finished" podID="776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" containerID="843d7a7fa9e2d4430c3baedec98c6e57411f55a0075a2d7e5799758eaadae500" exitCode=0 Dec 08 19:42:59 crc kubenswrapper[4706]: I1208 19:42:59.570639 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-brz4k" event={"ID":"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e","Type":"ContainerDied","Data":"843d7a7fa9e2d4430c3baedec98c6e57411f55a0075a2d7e5799758eaadae500"} Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.176900 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.239424 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-scripts\") pod \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.239557 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-certs\") pod \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.239722 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-config-data\") pod \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.239843 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78d8k\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-kube-api-access-78d8k\") pod \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.239899 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-combined-ca-bundle\") pod \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\" (UID: \"e68efceb-8175-4600-bc86-0d9d5d9d5dbd\") " Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.247533 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-certs" (OuterVolumeSpecName: "certs") pod "e68efceb-8175-4600-bc86-0d9d5d9d5dbd" (UID: "e68efceb-8175-4600-bc86-0d9d5d9d5dbd"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.250498 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-kube-api-access-78d8k" (OuterVolumeSpecName: "kube-api-access-78d8k") pod "e68efceb-8175-4600-bc86-0d9d5d9d5dbd" (UID: "e68efceb-8175-4600-bc86-0d9d5d9d5dbd"). InnerVolumeSpecName "kube-api-access-78d8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.268706 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-scripts" (OuterVolumeSpecName: "scripts") pod "e68efceb-8175-4600-bc86-0d9d5d9d5dbd" (UID: "e68efceb-8175-4600-bc86-0d9d5d9d5dbd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.303918 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-config-data" (OuterVolumeSpecName: "config-data") pod "e68efceb-8175-4600-bc86-0d9d5d9d5dbd" (UID: "e68efceb-8175-4600-bc86-0d9d5d9d5dbd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.312319 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e68efceb-8175-4600-bc86-0d9d5d9d5dbd" (UID: "e68efceb-8175-4600-bc86-0d9d5d9d5dbd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.342780 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.342820 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78d8k\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-kube-api-access-78d8k\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.342833 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.342842 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.342850 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e68efceb-8175-4600-bc86-0d9d5d9d5dbd-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.584149 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-59fls" event={"ID":"e68efceb-8175-4600-bc86-0d9d5d9d5dbd","Type":"ContainerDied","Data":"9daeb96ebafe0f18462455c507773e2b8569079f99a7b2f4372b5cc763a3a1e9"} Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.584238 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9daeb96ebafe0f18462455c507773e2b8569079f99a7b2f4372b5cc763a3a1e9" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.584174 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-59fls" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.873393 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-np9k8"] Dec 08 19:43:00 crc kubenswrapper[4706]: E1208 19:43:00.874376 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e68efceb-8175-4600-bc86-0d9d5d9d5dbd" containerName="cloudkitty-db-sync" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.874396 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e68efceb-8175-4600-bc86-0d9d5d9d5dbd" containerName="cloudkitty-db-sync" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.874612 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e68efceb-8175-4600-bc86-0d9d5d9d5dbd" containerName="cloudkitty-db-sync" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.875463 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.882942 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-fzwvl" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.883166 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.883380 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.883508 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.883633 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.911667 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-np9k8"] Dec 08 19:43:00 crc kubenswrapper[4706]: I1208 19:43:00.966613 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.051189 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k676\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-kube-api-access-7k676\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.051400 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-config-data\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.051500 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-certs\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.051529 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-combined-ca-bundle\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.051637 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-scripts\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.085087 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d88d7b95f-82ttp"] Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.090727 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" podUID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerName="dnsmasq-dns" containerID="cri-o://1be6dd51e03f7d43b1a1729d0c885697e581e860d2f114e8ec39e783fe37d36c" gracePeriod=10 Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.157343 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-scripts\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.157449 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k676\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-kube-api-access-7k676\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.157633 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-config-data\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.157728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-certs\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.157752 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-combined-ca-bundle\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.168625 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-config-data\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.174985 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-scripts\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.178100 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-certs\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.180507 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-combined-ca-bundle\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.203393 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k676\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-kube-api-access-7k676\") pod \"cloudkitty-storageinit-np9k8\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:01 crc kubenswrapper[4706]: I1208 19:43:01.204719 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:02 crc kubenswrapper[4706]: I1208 19:43:02.418356 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5b87bf4766-2rxxp" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 19:43:02 crc kubenswrapper[4706]: I1208 19:43:02.653552 4706 generic.go:334] "Generic (PLEG): container finished" podID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerID="1be6dd51e03f7d43b1a1729d0c885697e581e860d2f114e8ec39e783fe37d36c" exitCode=0 Dec 08 19:43:02 crc kubenswrapper[4706]: I1208 19:43:02.653611 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" event={"ID":"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce","Type":"ContainerDied","Data":"1be6dd51e03f7d43b1a1729d0c885697e581e860d2f114e8ec39e783fe37d36c"} Dec 08 19:43:02 crc kubenswrapper[4706]: I1208 19:43:02.949811 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:43:03 crc kubenswrapper[4706]: I1208 19:43:03.001914 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:43:03 crc kubenswrapper[4706]: I1208 19:43:03.081300 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" podUID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.163:5353: connect: connection refused" Dec 08 19:43:03 crc kubenswrapper[4706]: I1208 19:43:03.162846 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.687736 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8s7bd" event={"ID":"e18991e1-84d5-4e32-9a05-ad08f31cbd60","Type":"ContainerDied","Data":"7178f1388f4e200919070f5139b9bb0c4e4df82be32c11c9635aeac6c4da3a17"} Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.688150 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7178f1388f4e200919070f5139b9bb0c4e4df82be32c11c9635aeac6c4da3a17" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.779092 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8s7bd" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.865381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-db-sync-config-data\") pod \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.865611 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fv45\" (UniqueName: \"kubernetes.io/projected/e18991e1-84d5-4e32-9a05-ad08f31cbd60-kube-api-access-9fv45\") pod \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.865767 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-config-data\") pod \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.865822 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-combined-ca-bundle\") pod \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\" (UID: \"e18991e1-84d5-4e32-9a05-ad08f31cbd60\") " Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.878135 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e18991e1-84d5-4e32-9a05-ad08f31cbd60" (UID: "e18991e1-84d5-4e32-9a05-ad08f31cbd60"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.878216 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e18991e1-84d5-4e32-9a05-ad08f31cbd60-kube-api-access-9fv45" (OuterVolumeSpecName: "kube-api-access-9fv45") pod "e18991e1-84d5-4e32-9a05-ad08f31cbd60" (UID: "e18991e1-84d5-4e32-9a05-ad08f31cbd60"). InnerVolumeSpecName "kube-api-access-9fv45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.902987 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e18991e1-84d5-4e32-9a05-ad08f31cbd60" (UID: "e18991e1-84d5-4e32-9a05-ad08f31cbd60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.927014 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-config-data" (OuterVolumeSpecName: "config-data") pod "e18991e1-84d5-4e32-9a05-ad08f31cbd60" (UID: "e18991e1-84d5-4e32-9a05-ad08f31cbd60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.973218 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fv45\" (UniqueName: \"kubernetes.io/projected/e18991e1-84d5-4e32-9a05-ad08f31cbd60-kube-api-access-9fv45\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.973549 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.973632 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:04 crc kubenswrapper[4706]: I1208 19:43:04.973697 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e18991e1-84d5-4e32-9a05-ad08f31cbd60-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.637203 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-79bc9d977-gw6pk" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.706207 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8s7bd" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.710166 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-brz4k" event={"ID":"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e","Type":"ContainerDied","Data":"620c4841ad21c6859c2997a524fb76855866089caaf868ade99e023c4dd3ba57"} Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.710246 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="620c4841ad21c6859c2997a524fb76855866089caaf868ade99e023c4dd3ba57" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.751565 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bbb444974-g2ncm"] Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.751960 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bbb444974-g2ncm" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerName="neutron-api" containerID="cri-o://5e3995ca1f1359ced582de216ad8ebc70c7f73aed776be390e16bce54a023321" gracePeriod=30 Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.752305 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bbb444974-g2ncm" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerName="neutron-httpd" containerID="cri-o://4a8e5955c775d933fb66a0b03ba1f45d3189af279c759b6ad3ef4b069678bc77" gracePeriod=30 Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.829560 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-brz4k" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.835923 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.835969 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.836012 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.837739 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5cdb7c8ca56123f7a237303ad544e086627575cfd4a9196d909f15ed3eb4ca79"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.837914 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://5cdb7c8ca56123f7a237303ad544e086627575cfd4a9196d909f15ed3eb4ca79" gracePeriod=600 Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.902671 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-config-data\") pod \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.902810 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-etc-machine-id\") pod \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.902930 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-db-sync-config-data\") pod \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.902993 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6s9t\" (UniqueName: \"kubernetes.io/projected/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-kube-api-access-t6s9t\") pod \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.903429 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" (UID: "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.904543 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.914226 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" (UID: "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:05 crc kubenswrapper[4706]: I1208 19:43:05.926564 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-kube-api-access-t6s9t" (OuterVolumeSpecName: "kube-api-access-t6s9t") pod "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" (UID: "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e"). InnerVolumeSpecName "kube-api-access-t6s9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.006722 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-combined-ca-bundle\") pod \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.006811 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-scripts\") pod \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\" (UID: \"776cb4c2-52c5-404a-b3ca-c83f93dcbd0e\") " Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.007591 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.007615 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6s9t\" (UniqueName: \"kubernetes.io/projected/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-kube-api-access-t6s9t\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.029551 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-scripts" (OuterVolumeSpecName: "scripts") pod "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" (UID: "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.043101 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-config-data" (OuterVolumeSpecName: "config-data") pod "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" (UID: "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.120219 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.120290 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.143051 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" (UID: "776cb4c2-52c5-404a-b3ca-c83f93dcbd0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.233314 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.277722 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8r7st"] Dec 08 19:43:06 crc kubenswrapper[4706]: E1208 19:43:06.278497 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18991e1-84d5-4e32-9a05-ad08f31cbd60" containerName="glance-db-sync" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.278517 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18991e1-84d5-4e32-9a05-ad08f31cbd60" containerName="glance-db-sync" Dec 08 19:43:06 crc kubenswrapper[4706]: E1208 19:43:06.278545 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" containerName="cinder-db-sync" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.278555 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" containerName="cinder-db-sync" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.278859 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" containerName="cinder-db-sync" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.278894 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e18991e1-84d5-4e32-9a05-ad08f31cbd60" containerName="glance-db-sync" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.280585 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.295053 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8r7st"] Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.336871 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-config\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.336927 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-svc\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.337003 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.337065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.337126 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.337145 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm64k\" (UniqueName: \"kubernetes.io/projected/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-kube-api-access-mm64k\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.438815 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-config\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.438864 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-svc\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.438919 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.438962 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.439013 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.439031 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm64k\" (UniqueName: \"kubernetes.io/projected/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-kube-api-access-mm64k\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.440328 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-config\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.440803 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.440904 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.441493 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-svc\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.445201 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.480812 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm64k\" (UniqueName: \"kubernetes.io/projected/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-kube-api-access-mm64k\") pod \"dnsmasq-dns-85ff748b95-8r7st\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.646417 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.774383 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="5cdb7c8ca56123f7a237303ad544e086627575cfd4a9196d909f15ed3eb4ca79" exitCode=0 Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.774471 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"5cdb7c8ca56123f7a237303ad544e086627575cfd4a9196d909f15ed3eb4ca79"} Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.774555 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"a5681b5fe1d33f6e2441d03b7699cdb0e62dc0c63f422690f0cadd789557d42f"} Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.774585 4706 scope.go:117] "RemoveContainer" containerID="44273cc3a85c65641d37bbdee9218aa6e6f7747d6e135429b54fe362e247061f" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.794801 4706 generic.go:334] "Generic (PLEG): container finished" podID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerID="4a8e5955c775d933fb66a0b03ba1f45d3189af279c759b6ad3ef4b069678bc77" exitCode=0 Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.794911 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-brz4k" Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.794897 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbb444974-g2ncm" event={"ID":"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147","Type":"ContainerDied","Data":"4a8e5955c775d933fb66a0b03ba1f45d3189af279c759b6ad3ef4b069678bc77"} Dec 08 19:43:06 crc kubenswrapper[4706]: I1208 19:43:06.953169 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:43:07 crc kubenswrapper[4706]: E1208 19:43:07.024482 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.064757 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-nb\") pod \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.064893 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh7qm\" (UniqueName: \"kubernetes.io/projected/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-kube-api-access-bh7qm\") pod \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.066022 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-np9k8"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.067307 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-sb\") pod \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.067522 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-config\") pod \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.067577 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-swift-storage-0\") pod \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.067598 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-svc\") pod \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\" (UID: \"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce\") " Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.083659 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-kube-api-access-bh7qm" (OuterVolumeSpecName: "kube-api-access-bh7qm") pod "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" (UID: "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce"). InnerVolumeSpecName "kube-api-access-bh7qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.193615 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh7qm\" (UniqueName: \"kubernetes.io/projected/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-kube-api-access-bh7qm\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.258017 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" (UID: "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.275861 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" (UID: "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.337429 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.339085 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.339117 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:07 crc kubenswrapper[4706]: E1208 19:43:07.339999 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerName="init" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.340140 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerName="init" Dec 08 19:43:07 crc kubenswrapper[4706]: E1208 19:43:07.340302 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerName="dnsmasq-dns" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.340430 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerName="dnsmasq-dns" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.340749 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-config" (OuterVolumeSpecName: "config") pod "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" (UID: "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.341084 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" containerName="dnsmasq-dns" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.359808 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" (UID: "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.360701 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" (UID: "13b7ef6f-ece2-44fb-9a39-64ecfa9abbce"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.395174 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.406888 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kvdqs" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.407329 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.407529 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.419546 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.427591 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.442526 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.442573 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.442584 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.506625 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.521606 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.535958 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.536317 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.536371 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.536543 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-z4r6h" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.545085 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.545188 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-logs\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.545224 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.545314 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.545459 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smc24\" (UniqueName: \"kubernetes.io/projected/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-kube-api-access-smc24\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.545504 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.545535 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.566208 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647303 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647376 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp5bm\" (UniqueName: \"kubernetes.io/projected/a3505923-b491-491d-8c83-5a678f23c807-kube-api-access-tp5bm\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647413 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smc24\" (UniqueName: \"kubernetes.io/projected/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-kube-api-access-smc24\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647469 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647503 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647564 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3505923-b491-491d-8c83-5a678f23c807-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647591 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647624 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-logs\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647662 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647690 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647720 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647744 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.647785 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.648993 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.649254 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-logs\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.668189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.668965 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.682241 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.689761 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8r7st"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.690009 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.692323 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.700239 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smc24\" (UniqueName: \"kubernetes.io/projected/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-kube-api-access-smc24\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.706671 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.706760 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/67f02492c48f5b6b986fa4270f8226cd0203948331bc87047321fea71b95adc3/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.714427 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.761893 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-t2s6s"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.764364 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3505923-b491-491d-8c83-5a678f23c807-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.764466 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.764493 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.764530 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.764629 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.764675 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp5bm\" (UniqueName: \"kubernetes.io/projected/a3505923-b491-491d-8c83-5a678f23c807-kube-api-access-tp5bm\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.765298 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3505923-b491-491d-8c83-5a678f23c807-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.765422 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.784915 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.788117 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.788633 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.791613 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.792193 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.806103 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp5bm\" (UniqueName: \"kubernetes.io/projected/a3505923-b491-491d-8c83-5a678f23c807-kube-api-access-tp5bm\") pod \"cinder-scheduler-0\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.833770 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-t2s6s"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.873823 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.873888 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.873929 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.896529 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j72n\" (UniqueName: \"kubernetes.io/projected/6144d9b7-058e-47f2-b1f8-591554ec326c-kube-api-access-5j72n\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.896663 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.896720 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.896763 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.896783 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plbgf\" (UniqueName: \"kubernetes.io/projected/8fe0bc14-1790-4182-8c53-4e78f2c435b4-kube-api-access-plbgf\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.896841 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.897029 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.897108 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-config\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.897141 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.897205 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-logs\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.897804 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.902434 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8r7st"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.909214 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.911004 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a298368b-7b89-4c7a-8427-330e9d16ea00","Type":"ContainerStarted","Data":"f66aece7c66c7292b000dd869655d2fbda9db3a5c5cae845a60edd0464ed5fa9"} Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.911229 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="ceilometer-notification-agent" containerID="cri-o://6b59caf3b749aaf5fe740648eb1227d7639d865de8dba1b4f12c5c386f63dd51" gracePeriod=30 Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.911689 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.912416 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="proxy-httpd" containerID="cri-o://f66aece7c66c7292b000dd869655d2fbda9db3a5c5cae845a60edd0464ed5fa9" gracePeriod=30 Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.912481 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="sg-core" containerID="cri-o://a4a69a2c1c7767822f0dea5bbb310b47cc97d60c9ab969d5765ed5bb27f7edcc" gracePeriod=30 Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.941764 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.944347 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.946745 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 08 19:43:07 crc kubenswrapper[4706]: I1208 19:43:07.988036 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003661 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-config\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003718 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003755 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-logs\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003788 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-logs\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003813 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003845 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003866 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003892 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data-custom\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003935 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003963 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.003983 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.004035 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-scripts\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.004059 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j72n\" (UniqueName: \"kubernetes.io/projected/6144d9b7-058e-47f2-b1f8-591554ec326c-kube-api-access-5j72n\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.004096 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.004131 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.004162 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plbgf\" (UniqueName: \"kubernetes.io/projected/8fe0bc14-1790-4182-8c53-4e78f2c435b4-kube-api-access-plbgf\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.004178 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.004203 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.013405 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.014415 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.021134 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dhp6\" (UniqueName: \"kubernetes.io/projected/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-kube-api-access-9dhp6\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.019004 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.019449 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.019721 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-logs\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.020232 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.020435 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.017708 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-config\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.022204 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.058347 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.058405 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b040d665527e04c44fa620444bb13c4fde426c5d08f49faa6c04df54fad7a449/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.061369 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.062925 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.065129 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plbgf\" (UniqueName: \"kubernetes.io/projected/8fe0bc14-1790-4182-8c53-4e78f2c435b4-kube-api-access-plbgf\") pod \"dnsmasq-dns-5c9776ccc5-t2s6s\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.076895 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.099233 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j72n\" (UniqueName: \"kubernetes.io/projected/6144d9b7-058e-47f2-b1f8-591554ec326c-kube-api-access-5j72n\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.110005 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-8r7st" event={"ID":"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4","Type":"ContainerStarted","Data":"276439ba62d022dad5504d9533a36bc401b06bf437dd83b70080315ba700da22"} Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.124396 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dhp6\" (UniqueName: \"kubernetes.io/projected/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-kube-api-access-9dhp6\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.124464 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-logs\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.124530 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.124555 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data-custom\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.124591 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.124609 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.124649 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-scripts\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.126589 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-logs\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.127493 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.134703 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data-custom\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.135569 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.146234 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-np9k8" event={"ID":"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1","Type":"ContainerStarted","Data":"d6b9699ca260769714dd92bd847e10ad0349af7a1c4db2688c6afe4a9dd0a788"} Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.146314 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-np9k8" event={"ID":"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1","Type":"ContainerStarted","Data":"3d864f0c385615b5ebb76cf23fb175bd889359da98f770062eb146b069bc07f4"} Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.152872 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-scripts\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.166916 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dhp6\" (UniqueName: \"kubernetes.io/projected/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-kube-api-access-9dhp6\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.171594 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data\") pod \"cinder-api-0\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " pod="openstack/cinder-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.186912 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-np9k8" podStartSLOduration=8.186879018 podStartE2EDuration="8.186879018s" podCreationTimestamp="2025-12-08 19:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:08.171988921 +0000 UTC m=+1270.814189924" watchObservedRunningTime="2025-12-08 19:43:08.186879018 +0000 UTC m=+1270.829080021" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.187016 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.195168 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" event={"ID":"13b7ef6f-ece2-44fb-9a39-64ecfa9abbce","Type":"ContainerDied","Data":"050c3c79097bad4386b5fb14a9a2a41d6ee36c162afbd4d5f8c99da9809a3f36"} Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.195236 4706 scope.go:117] "RemoveContainer" containerID="1be6dd51e03f7d43b1a1729d0c885697e581e860d2f114e8ec39e783fe37d36c" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.195496 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d88d7b95f-82ttp" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.273994 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:08 crc kubenswrapper[4706]: I1208 19:43:08.322063 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.196248 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.220495 4706 scope.go:117] "RemoveContainer" containerID="b27f249bc9b1e5e681913b0b7145807cc1ff04cde87db165059ee5be72765df2" Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.303961 4706 generic.go:334] "Generic (PLEG): container finished" podID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerID="5e3995ca1f1359ced582de216ad8ebc70c7f73aed776be390e16bce54a023321" exitCode=0 Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.304136 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbb444974-g2ncm" event={"ID":"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147","Type":"ContainerDied","Data":"5e3995ca1f1359ced582de216ad8ebc70c7f73aed776be390e16bce54a023321"} Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.313344 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d88d7b95f-82ttp"] Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.324320 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8466877bdb-zbttb" Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.336083 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d88d7b95f-82ttp"] Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.480143 4706 generic.go:334] "Generic (PLEG): container finished" podID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerID="f66aece7c66c7292b000dd869655d2fbda9db3a5c5cae845a60edd0464ed5fa9" exitCode=0 Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.480192 4706 generic.go:334] "Generic (PLEG): container finished" podID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerID="a4a69a2c1c7767822f0dea5bbb310b47cc97d60c9ab969d5765ed5bb27f7edcc" exitCode=2 Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.480376 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a298368b-7b89-4c7a-8427-330e9d16ea00","Type":"ContainerDied","Data":"f66aece7c66c7292b000dd869655d2fbda9db3a5c5cae845a60edd0464ed5fa9"} Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.480444 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a298368b-7b89-4c7a-8427-330e9d16ea00","Type":"ContainerDied","Data":"a4a69a2c1c7767822f0dea5bbb310b47cc97d60c9ab969d5765ed5bb27f7edcc"} Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.547300 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5b87bf4766-2rxxp"] Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.548108 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5b87bf4766-2rxxp" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api-log" containerID="cri-o://d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f" gracePeriod=30 Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.549452 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5b87bf4766-2rxxp" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api" containerID="cri-o://2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec" gracePeriod=30 Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.559225 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b87bf4766-2rxxp" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": EOF" Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.676995 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13b7ef6f-ece2-44fb-9a39-64ecfa9abbce" path="/var/lib/kubelet/pods/13b7ef6f-ece2-44fb-9a39-64ecfa9abbce/volumes" Dec 08 19:43:09 crc kubenswrapper[4706]: I1208 19:43:09.938972 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.001864 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2gqz\" (UniqueName: \"kubernetes.io/projected/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-kube-api-access-g2gqz\") pod \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.002476 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-ovndb-tls-certs\") pod \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.002655 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-httpd-config\") pod \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.002733 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-config\") pod \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.002815 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-combined-ca-bundle\") pod \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\" (UID: \"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147\") " Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.023756 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-kube-api-access-g2gqz" (OuterVolumeSpecName: "kube-api-access-g2gqz") pod "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" (UID: "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147"). InnerVolumeSpecName "kube-api-access-g2gqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.031746 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" (UID: "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.115623 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-config" (OuterVolumeSpecName: "config") pod "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" (UID: "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.126085 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2gqz\" (UniqueName: \"kubernetes.io/projected/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-kube-api-access-g2gqz\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.126122 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.126132 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:10 crc kubenswrapper[4706]: E1208 19:43:10.150773 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37655fed_e807_4ad5_ada9_d2c4b0204e1d.slice/crio-d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37655fed_e807_4ad5_ada9_d2c4b0204e1d.slice/crio-conmon-d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda67d019f_ccc3_4cce_bf7c_96cc704f3ba4.slice/crio-conmon-7f760e00a07d19dc0266c990f689425b9d463a7e3b7c71ec454961b2866ff47d.scope\": RecentStats: unable to find data in memory cache]" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.174564 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" (UID: "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.215940 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" (UID: "05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.228306 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.228382 4706 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.372345 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.407683 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-t2s6s"] Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.483454 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.547365 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.552067 4706 generic.go:334] "Generic (PLEG): container finished" podID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerID="d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f" exitCode=143 Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.552190 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b87bf4766-2rxxp" event={"ID":"37655fed-e807-4ad5-ada9-d2c4b0204e1d","Type":"ContainerDied","Data":"d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f"} Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.558417 4706 generic.go:334] "Generic (PLEG): container finished" podID="a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" containerID="7f760e00a07d19dc0266c990f689425b9d463a7e3b7c71ec454961b2866ff47d" exitCode=0 Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.559469 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-8r7st" event={"ID":"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4","Type":"ContainerDied","Data":"7f760e00a07d19dc0266c990f689425b9d463a7e3b7c71ec454961b2866ff47d"} Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.719313 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbb444974-g2ncm" event={"ID":"05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147","Type":"ContainerDied","Data":"b732a5ae92086a1ffe077f23d24d56616cf2296b07c236ded6a8b9b2ae46b7d6"} Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.719385 4706 scope.go:117] "RemoveContainer" containerID="4a8e5955c775d933fb66a0b03ba1f45d3189af279c759b6ad3ef4b069678bc77" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.719639 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bbb444974-g2ncm" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.731389 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.826327 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.826826 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3505923-b491-491d-8c83-5a678f23c807","Type":"ContainerStarted","Data":"ddec2e7ee9bf9f2e6f7e542b5647395af42bbf3c94308d43dc25775cc203e381"} Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.862008 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" event={"ID":"8fe0bc14-1790-4182-8c53-4e78f2c435b4","Type":"ContainerStarted","Data":"3a0fe7298466080f637c0c8654e6576c8b2218d978f5195345082a2d04e99786"} Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.910477 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6144d9b7-058e-47f2-b1f8-591554ec326c","Type":"ContainerStarted","Data":"936e751a7212ecf93222aeb8332843da2c1b4b89a7196ae1355e42564a2022e5"} Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.937499 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.965596 4706 scope.go:117] "RemoveContainer" containerID="5e3995ca1f1359ced582de216ad8ebc70c7f73aed776be390e16bce54a023321" Dec 08 19:43:10 crc kubenswrapper[4706]: I1208 19:43:10.973483 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bbb444974-g2ncm"] Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.015114 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7bbb444974-g2ncm"] Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.639000 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" path="/var/lib/kubelet/pods/05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147/volumes" Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.847563 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.946279 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-8466877bdb-zbttb" podUID="b0acc09f-0610-4121-b42b-ebe7274b5c5d" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.172:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.953812 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-8466877bdb-zbttb" podUID="b0acc09f-0610-4121-b42b-ebe7274b5c5d" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.172:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.966161 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5","Type":"ContainerStarted","Data":"f338696d2a4c76b413a78064fa33550acc63acf41c3c1d2fe05ba16cb5f09208"} Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.998658 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-8r7st" event={"ID":"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4","Type":"ContainerDied","Data":"276439ba62d022dad5504d9533a36bc401b06bf437dd83b70080315ba700da22"} Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.998726 4706 scope.go:117] "RemoveContainer" containerID="7f760e00a07d19dc0266c990f689425b9d463a7e3b7c71ec454961b2866ff47d" Dec 08 19:43:11 crc kubenswrapper[4706]: I1208 19:43:11.998895 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-8r7st" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.010647 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.041159 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-sb\") pod \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.041289 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm64k\" (UniqueName: \"kubernetes.io/projected/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-kube-api-access-mm64k\") pod \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.041434 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-config\") pod \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.041567 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-svc\") pod \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.041693 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-nb\") pod \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.041718 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-swift-storage-0\") pod \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\" (UID: \"a67d019f-ccc3-4cce-bf7c-96cc704f3ba4\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.043560 4706 generic.go:334] "Generic (PLEG): container finished" podID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" containerID="308b1174f4d0dcacafe41dee6beebc2e3e8af97ca5387600777d14db5e00e238" exitCode=0 Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.043675 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" event={"ID":"8fe0bc14-1790-4182-8c53-4e78f2c435b4","Type":"ContainerDied","Data":"308b1174f4d0dcacafe41dee6beebc2e3e8af97ca5387600777d14db5e00e238"} Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.072322 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-kube-api-access-mm64k" (OuterVolumeSpecName: "kube-api-access-mm64k") pod "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" (UID: "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4"). InnerVolumeSpecName "kube-api-access-mm64k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.083541 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c","Type":"ContainerStarted","Data":"d47701fd0c7dfe4a7a55eb956dc8d839d6d14fa4ffed74d0884074c444cbf352"} Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.135313 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" (UID: "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.145061 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-combined-ca-bundle\") pod \"a298368b-7b89-4c7a-8427-330e9d16ea00\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.145416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-sg-core-conf-yaml\") pod \"a298368b-7b89-4c7a-8427-330e9d16ea00\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.145492 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-scripts\") pod \"a298368b-7b89-4c7a-8427-330e9d16ea00\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.145568 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-log-httpd\") pod \"a298368b-7b89-4c7a-8427-330e9d16ea00\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.145600 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-config-data\") pod \"a298368b-7b89-4c7a-8427-330e9d16ea00\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.145636 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-run-httpd\") pod \"a298368b-7b89-4c7a-8427-330e9d16ea00\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.145767 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g265b\" (UniqueName: \"kubernetes.io/projected/a298368b-7b89-4c7a-8427-330e9d16ea00-kube-api-access-g265b\") pod \"a298368b-7b89-4c7a-8427-330e9d16ea00\" (UID: \"a298368b-7b89-4c7a-8427-330e9d16ea00\") " Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.146876 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.146894 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm64k\" (UniqueName: \"kubernetes.io/projected/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-kube-api-access-mm64k\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.153426 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a298368b-7b89-4c7a-8427-330e9d16ea00" (UID: "a298368b-7b89-4c7a-8427-330e9d16ea00"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.154187 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a298368b-7b89-4c7a-8427-330e9d16ea00" (UID: "a298368b-7b89-4c7a-8427-330e9d16ea00"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.157119 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" (UID: "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.165073 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a298368b-7b89-4c7a-8427-330e9d16ea00-kube-api-access-g265b" (OuterVolumeSpecName: "kube-api-access-g265b") pod "a298368b-7b89-4c7a-8427-330e9d16ea00" (UID: "a298368b-7b89-4c7a-8427-330e9d16ea00"). InnerVolumeSpecName "kube-api-access-g265b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.172591 4706 generic.go:334] "Generic (PLEG): container finished" podID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerID="6b59caf3b749aaf5fe740648eb1227d7639d865de8dba1b4f12c5c386f63dd51" exitCode=0 Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.173330 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a298368b-7b89-4c7a-8427-330e9d16ea00","Type":"ContainerDied","Data":"6b59caf3b749aaf5fe740648eb1227d7639d865de8dba1b4f12c5c386f63dd51"} Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.173509 4706 scope.go:117] "RemoveContainer" containerID="f66aece7c66c7292b000dd869655d2fbda9db3a5c5cae845a60edd0464ed5fa9" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.173815 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.206469 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-scripts" (OuterVolumeSpecName: "scripts") pod "a298368b-7b89-4c7a-8427-330e9d16ea00" (UID: "a298368b-7b89-4c7a-8427-330e9d16ea00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.235014 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-config" (OuterVolumeSpecName: "config") pod "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" (UID: "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.249170 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.249547 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.249634 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a298368b-7b89-4c7a-8427-330e9d16ea00-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.249756 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.249843 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g265b\" (UniqueName: \"kubernetes.io/projected/a298368b-7b89-4c7a-8427-330e9d16ea00-kube-api-access-g265b\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.249920 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.255019 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" (UID: "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.277767 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" (UID: "a67d019f-ccc3-4cce-bf7c-96cc704f3ba4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.298514 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a298368b-7b89-4c7a-8427-330e9d16ea00" (UID: "a298368b-7b89-4c7a-8427-330e9d16ea00"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.353701 4706 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.353753 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.353772 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.385482 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a298368b-7b89-4c7a-8427-330e9d16ea00" (UID: "a298368b-7b89-4c7a-8427-330e9d16ea00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.386287 4706 scope.go:117] "RemoveContainer" containerID="a4a69a2c1c7767822f0dea5bbb310b47cc97d60c9ab969d5765ed5bb27f7edcc" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.440519 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-config-data" (OuterVolumeSpecName: "config-data") pod "a298368b-7b89-4c7a-8427-330e9d16ea00" (UID: "a298368b-7b89-4c7a-8427-330e9d16ea00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.456112 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.456171 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a298368b-7b89-4c7a-8427-330e9d16ea00-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.624552 4706 scope.go:117] "RemoveContainer" containerID="6b59caf3b749aaf5fe740648eb1227d7639d865de8dba1b4f12c5c386f63dd51" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.764798 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.808532 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.834912 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.861183 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:12 crc kubenswrapper[4706]: E1208 19:43:12.861821 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerName="neutron-api" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.861843 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerName="neutron-api" Dec 08 19:43:12 crc kubenswrapper[4706]: E1208 19:43:12.861970 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerName="neutron-httpd" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.861982 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerName="neutron-httpd" Dec 08 19:43:12 crc kubenswrapper[4706]: E1208 19:43:12.861989 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="sg-core" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.861996 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="sg-core" Dec 08 19:43:12 crc kubenswrapper[4706]: E1208 19:43:12.862051 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="proxy-httpd" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862058 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="proxy-httpd" Dec 08 19:43:12 crc kubenswrapper[4706]: E1208 19:43:12.862070 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="ceilometer-notification-agent" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862077 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="ceilometer-notification-agent" Dec 08 19:43:12 crc kubenswrapper[4706]: E1208 19:43:12.862106 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" containerName="init" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862113 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" containerName="init" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862460 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerName="neutron-httpd" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862473 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" containerName="init" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862511 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="05f612d9-fc1c-4dcd-9ccf-9eb7c41fd147" containerName="neutron-api" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862526 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="ceilometer-notification-agent" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862538 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="proxy-httpd" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.862545 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" containerName="sg-core" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.866155 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.876145 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.876578 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.881014 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8r7st"] Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.911694 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.930866 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8r7st"] Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.995663 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-log-httpd\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.995866 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-scripts\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.995902 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.996173 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-run-httpd\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.996853 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.997023 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66ls4\" (UniqueName: \"kubernetes.io/projected/f9f63de4-e7de-4a25-8044-0c2603442785-kube-api-access-66ls4\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:12 crc kubenswrapper[4706]: I1208 19:43:12.997337 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-config-data\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.103953 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.104023 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66ls4\" (UniqueName: \"kubernetes.io/projected/f9f63de4-e7de-4a25-8044-0c2603442785-kube-api-access-66ls4\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.104059 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-config-data\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.104127 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-log-httpd\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.106338 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-log-httpd\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.106690 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-scripts\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.106713 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.106767 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-run-httpd\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.107149 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-run-httpd\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.122287 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.122993 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-scripts\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.128909 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.138745 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66ls4\" (UniqueName: \"kubernetes.io/projected/f9f63de4-e7de-4a25-8044-0c2603442785-kube-api-access-66ls4\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.149362 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-config-data\") pod \"ceilometer-0\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.208669 4706 generic.go:334] "Generic (PLEG): container finished" podID="2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" containerID="d6b9699ca260769714dd92bd847e10ad0349af7a1c4db2688c6afe4a9dd0a788" exitCode=0 Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.208754 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-np9k8" event={"ID":"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1","Type":"ContainerDied","Data":"d6b9699ca260769714dd92bd847e10ad0349af7a1c4db2688c6afe4a9dd0a788"} Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.218002 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.226671 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" event={"ID":"8fe0bc14-1790-4182-8c53-4e78f2c435b4","Type":"ContainerStarted","Data":"308973cbb8adda8e5cbccd2e780ac364c5aa1e4179a6c79616b71a25f3d27c80"} Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.227920 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.238127 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6144d9b7-058e-47f2-b1f8-591554ec326c","Type":"ContainerStarted","Data":"5a0d1e58d86280e31f33a9872063c705e55c8630b63ee2b818e7cba9c38d78e4"} Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.243971 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c","Type":"ContainerStarted","Data":"7ba2d2cc043b5e2d0aba651d7b6f4524658bde9741005f5bf327fe502ec52595"} Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.263851 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" podStartSLOduration=6.263833472 podStartE2EDuration="6.263833472s" podCreationTimestamp="2025-12-08 19:43:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:13.263603885 +0000 UTC m=+1275.905804888" watchObservedRunningTime="2025-12-08 19:43:13.263833472 +0000 UTC m=+1275.906034475" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.267138 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5","Type":"ContainerStarted","Data":"5735209d3c1edb50e1e06048cd3b2f42231803b94b7891a100e7e1e729bf90ad"} Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.635870 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a298368b-7b89-4c7a-8427-330e9d16ea00" path="/var/lib/kubelet/pods/a298368b-7b89-4c7a-8427-330e9d16ea00/volumes" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.638003 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a67d019f-ccc3-4cce-bf7c-96cc704f3ba4" path="/var/lib/kubelet/pods/a67d019f-ccc3-4cce-bf7c-96cc704f3ba4/volumes" Dec 08 19:43:13 crc kubenswrapper[4706]: I1208 19:43:13.930443 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:14 crc kubenswrapper[4706]: I1208 19:43:14.276041 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b87bf4766-2rxxp" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": read tcp 10.217.0.2:43444->10.217.0.169:9311: read: connection reset by peer" Dec 08 19:43:14 crc kubenswrapper[4706]: I1208 19:43:14.278954 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b87bf4766-2rxxp" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": read tcp 10.217.0.2:43428->10.217.0.169:9311: read: connection reset by peer" Dec 08 19:43:14 crc kubenswrapper[4706]: I1208 19:43:14.297934 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3505923-b491-491d-8c83-5a678f23c807","Type":"ContainerStarted","Data":"4dcc8bfad25c9e3360be737cbd4c1a87e35a9121184b4f42357a416193b9a192"} Dec 08 19:43:14 crc kubenswrapper[4706]: I1208 19:43:14.306424 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerStarted","Data":"ac6c4653f38bb585725a7a40dcf9155b061696baafcf1c22c04ecad1780c92d1"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.003130 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.081027 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-scripts\") pod \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.081075 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-config-data\") pod \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.081171 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-certs\") pod \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.081336 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k676\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-kube-api-access-7k676\") pod \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.081543 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-combined-ca-bundle\") pod \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\" (UID: \"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.101060 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-certs" (OuterVolumeSpecName: "certs") pod "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" (UID: "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.108467 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-scripts" (OuterVolumeSpecName: "scripts") pod "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" (UID: "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.114168 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.120053 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-kube-api-access-7k676" (OuterVolumeSpecName: "kube-api-access-7k676") pod "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" (UID: "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1"). InnerVolumeSpecName "kube-api-access-7k676". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.137890 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" (UID: "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.144531 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-config-data" (OuterVolumeSpecName: "config-data") pod "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" (UID: "2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.184243 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data-custom\") pod \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.184576 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37655fed-e807-4ad5-ada9-d2c4b0204e1d-logs\") pod \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.184702 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data\") pod \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.184762 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-combined-ca-bundle\") pod \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.184792 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckhr6\" (UniqueName: \"kubernetes.io/projected/37655fed-e807-4ad5-ada9-d2c4b0204e1d-kube-api-access-ckhr6\") pod \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\" (UID: \"37655fed-e807-4ad5-ada9-d2c4b0204e1d\") " Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.186256 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37655fed-e807-4ad5-ada9-d2c4b0204e1d-logs" (OuterVolumeSpecName: "logs") pod "37655fed-e807-4ad5-ada9-d2c4b0204e1d" (UID: "37655fed-e807-4ad5-ada9-d2c4b0204e1d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.186629 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.186645 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k676\" (UniqueName: \"kubernetes.io/projected/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-kube-api-access-7k676\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.186654 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.186663 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37655fed-e807-4ad5-ada9-d2c4b0204e1d-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.186671 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.186680 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.190950 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "37655fed-e807-4ad5-ada9-d2c4b0204e1d" (UID: "37655fed-e807-4ad5-ada9-d2c4b0204e1d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.198670 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37655fed-e807-4ad5-ada9-d2c4b0204e1d-kube-api-access-ckhr6" (OuterVolumeSpecName: "kube-api-access-ckhr6") pod "37655fed-e807-4ad5-ada9-d2c4b0204e1d" (UID: "37655fed-e807-4ad5-ada9-d2c4b0204e1d"). InnerVolumeSpecName "kube-api-access-ckhr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.220511 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37655fed-e807-4ad5-ada9-d2c4b0204e1d" (UID: "37655fed-e807-4ad5-ada9-d2c4b0204e1d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.245000 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data" (OuterVolumeSpecName: "config-data") pod "37655fed-e807-4ad5-ada9-d2c4b0204e1d" (UID: "37655fed-e807-4ad5-ada9-d2c4b0204e1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.289509 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.289561 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.289577 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckhr6\" (UniqueName: \"kubernetes.io/projected/37655fed-e807-4ad5-ada9-d2c4b0204e1d-kube-api-access-ckhr6\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.289587 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37655fed-e807-4ad5-ada9-d2c4b0204e1d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.326419 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6144d9b7-058e-47f2-b1f8-591554ec326c","Type":"ContainerStarted","Data":"b653d8ef209dfae850503ed2cfae03cd0af1c63334d4b8dcccac8c55d6f7140e"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.326661 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerName="glance-log" containerID="cri-o://5a0d1e58d86280e31f33a9872063c705e55c8630b63ee2b818e7cba9c38d78e4" gracePeriod=30 Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.327417 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerName="glance-httpd" containerID="cri-o://b653d8ef209dfae850503ed2cfae03cd0af1c63334d4b8dcccac8c55d6f7140e" gracePeriod=30 Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.332028 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerStarted","Data":"db5da6a4db8a9b74a63d29235c313b87049a39af1a933e120a14a75ccc453a05"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.335986 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c","Type":"ContainerStarted","Data":"4a415fe4709f1fa7a78e614f3b8db4232e11c0f2ebb2390d0d9d350dab6bc86f"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.336178 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerName="glance-log" containerID="cri-o://7ba2d2cc043b5e2d0aba651d7b6f4524658bde9741005f5bf327fe502ec52595" gracePeriod=30 Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.336311 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerName="glance-httpd" containerID="cri-o://4a415fe4709f1fa7a78e614f3b8db4232e11c0f2ebb2390d0d9d350dab6bc86f" gracePeriod=30 Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.346817 4706 generic.go:334] "Generic (PLEG): container finished" podID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerID="2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec" exitCode=0 Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.346878 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b87bf4766-2rxxp" event={"ID":"37655fed-e807-4ad5-ada9-d2c4b0204e1d","Type":"ContainerDied","Data":"2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.347325 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b87bf4766-2rxxp" event={"ID":"37655fed-e807-4ad5-ada9-d2c4b0204e1d","Type":"ContainerDied","Data":"43ed7cafb068d3ff39d33d3e38c438605c0865977a4f20c12f31eb5a8dda96cd"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.347397 4706 scope.go:117] "RemoveContainer" containerID="2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.349447 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b87bf4766-2rxxp" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.365648 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5","Type":"ContainerStarted","Data":"0081e07dcfb08df7141d07009610f066e1d284b0fae5b318fe562d0492bc804c"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.365645 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.365609902 podStartE2EDuration="9.365609902s" podCreationTimestamp="2025-12-08 19:43:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:15.359127776 +0000 UTC m=+1278.001328779" watchObservedRunningTime="2025-12-08 19:43:15.365609902 +0000 UTC m=+1278.007810935" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.366393 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerName="cinder-api" containerID="cri-o://0081e07dcfb08df7141d07009610f066e1d284b0fae5b318fe562d0492bc804c" gracePeriod=30 Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.366423 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.366001 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerName="cinder-api-log" containerID="cri-o://5735209d3c1edb50e1e06048cd3b2f42231803b94b7891a100e7e1e729bf90ad" gracePeriod=30 Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.382108 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3505923-b491-491d-8c83-5a678f23c807","Type":"ContainerStarted","Data":"113e4502b6e01a73f9bb145bee24b9906afb3a9a56d04bfafd2c84bda06eacf0"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.411613 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-np9k8" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.417410 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-np9k8" event={"ID":"2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1","Type":"ContainerDied","Data":"3d864f0c385615b5ebb76cf23fb175bd889359da98f770062eb146b069bc07f4"} Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.417636 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d864f0c385615b5ebb76cf23fb175bd889359da98f770062eb146b069bc07f4" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.453768 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.453733236 podStartE2EDuration="9.453733236s" podCreationTimestamp="2025-12-08 19:43:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:15.385042868 +0000 UTC m=+1278.027243881" watchObservedRunningTime="2025-12-08 19:43:15.453733236 +0000 UTC m=+1278.095934239" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.454323 4706 scope.go:117] "RemoveContainer" containerID="d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.465824 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=8.465795121 podStartE2EDuration="8.465795121s" podCreationTimestamp="2025-12-08 19:43:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:15.430172181 +0000 UTC m=+1278.072373184" watchObservedRunningTime="2025-12-08 19:43:15.465795121 +0000 UTC m=+1278.107996124" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.519589 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.122648816 podStartE2EDuration="8.519560641s" podCreationTimestamp="2025-12-08 19:43:07 +0000 UTC" firstStartedPulling="2025-12-08 19:43:10.4112853 +0000 UTC m=+1273.053486303" lastFinishedPulling="2025-12-08 19:43:11.808197135 +0000 UTC m=+1274.450398128" observedRunningTime="2025-12-08 19:43:15.497501749 +0000 UTC m=+1278.139702752" watchObservedRunningTime="2025-12-08 19:43:15.519560641 +0000 UTC m=+1278.161761644" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.728866 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.747014 4706 scope.go:117] "RemoveContainer" containerID="2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec" Dec 08 19:43:15 crc kubenswrapper[4706]: E1208 19:43:15.765008 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec\": container with ID starting with 2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec not found: ID does not exist" containerID="2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.765332 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec"} err="failed to get container status \"2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec\": rpc error: code = NotFound desc = could not find container \"2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec\": container with ID starting with 2a4a863f25f5b3b889ec5daa63b39301830de6af311b55730ad2da7efd22dbec not found: ID does not exist" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.765424 4706 scope.go:117] "RemoveContainer" containerID="d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f" Dec 08 19:43:15 crc kubenswrapper[4706]: E1208 19:43:15.765756 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" containerName="cloudkitty-storageinit" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.765827 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" containerName="cloudkitty-storageinit" Dec 08 19:43:15 crc kubenswrapper[4706]: E1208 19:43:15.765968 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api-log" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.766027 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api-log" Dec 08 19:43:15 crc kubenswrapper[4706]: E1208 19:43:15.766091 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.766177 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.766714 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api-log" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.766795 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" containerName="barbican-api" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.766853 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" containerName="cloudkitty-storageinit" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.769276 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: E1208 19:43:15.778483 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f\": container with ID starting with d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f not found: ID does not exist" containerID="d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.778533 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f"} err="failed to get container status \"d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f\": rpc error: code = NotFound desc = could not find container \"d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f\": container with ID starting with d233eed837963a664ff6ba17612bd187aef96586b001818ceb2b803ff3c7448f not found: ID does not exist" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.780703 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.781105 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-fzwvl" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.781541 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.782182 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.785693 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.791036 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.823836 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.823911 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrp9r\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-kube-api-access-xrp9r\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.823942 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.824046 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-scripts\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.824146 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.824278 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-certs\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.826782 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5b87bf4766-2rxxp"] Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.904367 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5b87bf4766-2rxxp"] Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.927215 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.927663 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrp9r\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-kube-api-access-xrp9r\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.927699 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.927807 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-scripts\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.927906 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.928015 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-certs\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.941702 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-t2s6s"] Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.949105 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.951318 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-scripts\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.952690 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.963802 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-certs\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.970619 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-bd49n"] Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.982855 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.986995 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrp9r\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-kube-api-access-xrp9r\") pod \"cloudkitty-proc-0\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.988288 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:15 crc kubenswrapper[4706]: I1208 19:43:15.988635 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-bd49n"] Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.027994 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.032072 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.032124 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-config\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.032175 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-svc\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.032209 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.032228 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msfx2\" (UniqueName: \"kubernetes.io/projected/0fe2563a-a800-40b6-84e6-60249516049f-kube-api-access-msfx2\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.032679 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.035891 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.049869 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.060351 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.139848 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.139949 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.139978 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq426\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-kube-api-access-pq426\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140014 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-config\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140047 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-scripts\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140106 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-svc\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140155 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140178 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msfx2\" (UniqueName: \"kubernetes.io/projected/0fe2563a-a800-40b6-84e6-60249516049f-kube-api-access-msfx2\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140223 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cf99b6-8613-4e30-97f4-99317f5780d8-logs\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140324 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140392 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140425 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.140448 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-certs\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.142095 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.142923 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.143500 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.143783 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-config\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.144250 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-svc\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.170562 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.179485 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msfx2\" (UniqueName: \"kubernetes.io/projected/0fe2563a-a800-40b6-84e6-60249516049f-kube-api-access-msfx2\") pod \"dnsmasq-dns-67bdc55879-bd49n\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.246281 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cf99b6-8613-4e30-97f4-99317f5780d8-logs\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.246925 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.247027 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.247050 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-certs\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.247155 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.247239 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq426\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-kube-api-access-pq426\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.247321 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-scripts\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.250030 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cf99b6-8613-4e30-97f4-99317f5780d8-logs\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.258185 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-certs\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.259085 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.273157 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.275015 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.279705 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq426\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-kube-api-access-pq426\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.313597 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-scripts\") pod \"cloudkitty-api-0\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.454047 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.467552 4706 generic.go:334] "Generic (PLEG): container finished" podID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerID="0081e07dcfb08df7141d07009610f066e1d284b0fae5b318fe562d0492bc804c" exitCode=0 Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.467602 4706 generic.go:334] "Generic (PLEG): container finished" podID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerID="5735209d3c1edb50e1e06048cd3b2f42231803b94b7891a100e7e1e729bf90ad" exitCode=143 Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.467724 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5","Type":"ContainerDied","Data":"0081e07dcfb08df7141d07009610f066e1d284b0fae5b318fe562d0492bc804c"} Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.467782 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5","Type":"ContainerDied","Data":"5735209d3c1edb50e1e06048cd3b2f42231803b94b7891a100e7e1e729bf90ad"} Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.474861 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.482815 4706 generic.go:334] "Generic (PLEG): container finished" podID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerID="b653d8ef209dfae850503ed2cfae03cd0af1c63334d4b8dcccac8c55d6f7140e" exitCode=0 Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.482853 4706 generic.go:334] "Generic (PLEG): container finished" podID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerID="5a0d1e58d86280e31f33a9872063c705e55c8630b63ee2b818e7cba9c38d78e4" exitCode=143 Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.482910 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6144d9b7-058e-47f2-b1f8-591554ec326c","Type":"ContainerDied","Data":"b653d8ef209dfae850503ed2cfae03cd0af1c63334d4b8dcccac8c55d6f7140e"} Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.482989 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6144d9b7-058e-47f2-b1f8-591554ec326c","Type":"ContainerDied","Data":"5a0d1e58d86280e31f33a9872063c705e55c8630b63ee2b818e7cba9c38d78e4"} Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.485515 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerStarted","Data":"d40db7467f3e1069452c5f73a15afa51bfda2da03ce4562dd1e2daa5a4b1c646"} Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.493575 4706 generic.go:334] "Generic (PLEG): container finished" podID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerID="4a415fe4709f1fa7a78e614f3b8db4232e11c0f2ebb2390d0d9d350dab6bc86f" exitCode=0 Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.493624 4706 generic.go:334] "Generic (PLEG): container finished" podID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerID="7ba2d2cc043b5e2d0aba651d7b6f4524658bde9741005f5bf327fe502ec52595" exitCode=143 Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.493650 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c","Type":"ContainerDied","Data":"4a415fe4709f1fa7a78e614f3b8db4232e11c0f2ebb2390d0d9d350dab6bc86f"} Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.493712 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c","Type":"ContainerDied","Data":"7ba2d2cc043b5e2d0aba651d7b6f4524658bde9741005f5bf327fe502ec52595"} Dec 08 19:43:16 crc kubenswrapper[4706]: I1208 19:43:16.494093 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" podUID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" containerName="dnsmasq-dns" containerID="cri-o://308973cbb8adda8e5cbccd2e780ac364c5aa1e4179a6c79616b71a25f3d27c80" gracePeriod=10 Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.180391 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.266866 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.305176 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-logs\") pod \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.305312 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data\") pod \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.305416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-combined-ca-bundle\") pod \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.305629 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data-custom\") pod \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.305652 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-scripts\") pod \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.305692 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dhp6\" (UniqueName: \"kubernetes.io/projected/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-kube-api-access-9dhp6\") pod \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.305761 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-etc-machine-id\") pod \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.306746 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" (UID: "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.307410 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-logs" (OuterVolumeSpecName: "logs") pod "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" (UID: "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.315402 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" (UID: "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.315573 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-kube-api-access-9dhp6" (OuterVolumeSpecName: "kube-api-access-9dhp6") pod "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" (UID: "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"). InnerVolumeSpecName "kube-api-access-9dhp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.322413 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-scripts" (OuterVolumeSpecName: "scripts") pod "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" (UID: "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.361763 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" (UID: "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.408432 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-config-data\") pod \"6144d9b7-058e-47f2-b1f8-591554ec326c\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.408515 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j72n\" (UniqueName: \"kubernetes.io/projected/6144d9b7-058e-47f2-b1f8-591554ec326c-kube-api-access-5j72n\") pod \"6144d9b7-058e-47f2-b1f8-591554ec326c\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.408683 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-combined-ca-bundle\") pod \"6144d9b7-058e-47f2-b1f8-591554ec326c\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.408736 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-httpd-run\") pod \"6144d9b7-058e-47f2-b1f8-591554ec326c\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.409130 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"6144d9b7-058e-47f2-b1f8-591554ec326c\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.409231 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-scripts\") pod \"6144d9b7-058e-47f2-b1f8-591554ec326c\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.409281 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-logs\") pod \"6144d9b7-058e-47f2-b1f8-591554ec326c\" (UID: \"6144d9b7-058e-47f2-b1f8-591554ec326c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.410167 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.410199 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.410216 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.410227 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.410239 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dhp6\" (UniqueName: \"kubernetes.io/projected/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-kube-api-access-9dhp6\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.410251 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.411538 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6144d9b7-058e-47f2-b1f8-591554ec326c" (UID: "6144d9b7-058e-47f2-b1f8-591554ec326c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.419750 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6144d9b7-058e-47f2-b1f8-591554ec326c-kube-api-access-5j72n" (OuterVolumeSpecName: "kube-api-access-5j72n") pod "6144d9b7-058e-47f2-b1f8-591554ec326c" (UID: "6144d9b7-058e-47f2-b1f8-591554ec326c"). InnerVolumeSpecName "kube-api-access-5j72n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.426466 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-logs" (OuterVolumeSpecName: "logs") pod "6144d9b7-058e-47f2-b1f8-591554ec326c" (UID: "6144d9b7-058e-47f2-b1f8-591554ec326c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.447630 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-scripts" (OuterVolumeSpecName: "scripts") pod "6144d9b7-058e-47f2-b1f8-591554ec326c" (UID: "6144d9b7-058e-47f2-b1f8-591554ec326c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.472826 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7" (OuterVolumeSpecName: "glance") pod "6144d9b7-058e-47f2-b1f8-591554ec326c" (UID: "6144d9b7-058e-47f2-b1f8-591554ec326c"). InnerVolumeSpecName "pvc-157f530e-9886-42dc-a031-b8a66480e0f7". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.498900 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6144d9b7-058e-47f2-b1f8-591554ec326c" (UID: "6144d9b7-058e-47f2-b1f8-591554ec326c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.515900 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data" (OuterVolumeSpecName: "config-data") pod "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" (UID: "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.516421 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data\") pod \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\" (UID: \"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5\") " Dec 08 19:43:17 crc kubenswrapper[4706]: W1208 19:43:17.516572 4706 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5/volumes/kubernetes.io~secret/config-data Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.516593 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data" (OuterVolumeSpecName: "config-data") pod "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" (UID: "42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.517350 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.517382 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.517396 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.517446 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") on node \"crc\" " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.517460 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.517472 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6144d9b7-058e-47f2-b1f8-591554ec326c-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.517483 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j72n\" (UniqueName: \"kubernetes.io/projected/6144d9b7-058e-47f2-b1f8-591554ec326c-kube-api-access-5j72n\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.530771 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerStarted","Data":"767012e581f3c74aeb46a9125c724ba78c8da18be184f6f10983a171033b3e12"} Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.545953 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"42efdc0a-1b75-4412-8e3f-a2ece0aaafe5","Type":"ContainerDied","Data":"f338696d2a4c76b413a78064fa33550acc63acf41c3c1d2fe05ba16cb5f09208"} Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.546019 4706 scope.go:117] "RemoveContainer" containerID="0081e07dcfb08df7141d07009610f066e1d284b0fae5b318fe562d0492bc804c" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.546148 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.569475 4706 generic.go:334] "Generic (PLEG): container finished" podID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" containerID="308973cbb8adda8e5cbccd2e780ac364c5aa1e4179a6c79616b71a25f3d27c80" exitCode=0 Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.569794 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" event={"ID":"8fe0bc14-1790-4182-8c53-4e78f2c435b4","Type":"ContainerDied","Data":"308973cbb8adda8e5cbccd2e780ac364c5aa1e4179a6c79616b71a25f3d27c80"} Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.577239 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6144d9b7-058e-47f2-b1f8-591554ec326c","Type":"ContainerDied","Data":"936e751a7212ecf93222aeb8332843da2c1b4b89a7196ae1355e42564a2022e5"} Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.577388 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.598928 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-config-data" (OuterVolumeSpecName: "config-data") pod "6144d9b7-058e-47f2-b1f8-591554ec326c" (UID: "6144d9b7-058e-47f2-b1f8-591554ec326c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.600525 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.601192 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-157f530e-9886-42dc-a031-b8a66480e0f7" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7") on node "crc" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.621657 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.621716 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6144d9b7-058e-47f2-b1f8-591554ec326c-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.626142 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37655fed-e807-4ad5-ada9-d2c4b0204e1d" path="/var/lib/kubelet/pods/37655fed-e807-4ad5-ada9-d2c4b0204e1d/volumes" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.697043 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.705803 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.732537 4706 scope.go:117] "RemoveContainer" containerID="5735209d3c1edb50e1e06048cd3b2f42231803b94b7891a100e7e1e729bf90ad" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.776254 4706 scope.go:117] "RemoveContainer" containerID="b653d8ef209dfae850503ed2cfae03cd0af1c63334d4b8dcccac8c55d6f7140e" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.806668 4706 scope.go:117] "RemoveContainer" containerID="5a0d1e58d86280e31f33a9872063c705e55c8630b63ee2b818e7cba9c38d78e4" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.836935 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-config-data\") pod \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837019 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-combined-ca-bundle\") pod \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837217 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837273 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-swift-storage-0\") pod \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837364 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-httpd-run\") pod \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837478 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-sb\") pod \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837572 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smc24\" (UniqueName: \"kubernetes.io/projected/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-kube-api-access-smc24\") pod \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837603 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-logs\") pod \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837690 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-config\") pod \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837743 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-svc\") pod \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837774 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-nb\") pod \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837803 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plbgf\" (UniqueName: \"kubernetes.io/projected/8fe0bc14-1790-4182-8c53-4e78f2c435b4-kube-api-access-plbgf\") pod \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\" (UID: \"8fe0bc14-1790-4182-8c53-4e78f2c435b4\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.837834 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-scripts\") pod \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\" (UID: \"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c\") " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.839222 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-logs" (OuterVolumeSpecName: "logs") pod "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" (UID: "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.839336 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" (UID: "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.844446 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-scripts" (OuterVolumeSpecName: "scripts") pod "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" (UID: "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.853755 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe0bc14-1790-4182-8c53-4e78f2c435b4-kube-api-access-plbgf" (OuterVolumeSpecName: "kube-api-access-plbgf") pod "8fe0bc14-1790-4182-8c53-4e78f2c435b4" (UID: "8fe0bc14-1790-4182-8c53-4e78f2c435b4"). InnerVolumeSpecName "kube-api-access-plbgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.859387 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-kube-api-access-smc24" (OuterVolumeSpecName: "kube-api-access-smc24") pod "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" (UID: "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c"). InnerVolumeSpecName "kube-api-access-smc24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.873324 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028" (OuterVolumeSpecName: "glance") pod "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" (UID: "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c"). InnerVolumeSpecName "pvc-de119cd6-b81e-445f-a9db-62090e36b028". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.899917 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.918894 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" (UID: "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.922051 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8fe0bc14-1790-4182-8c53-4e78f2c435b4" (UID: "8fe0bc14-1790-4182-8c53-4e78f2c435b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.934446 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-config" (OuterVolumeSpecName: "config") pod "8fe0bc14-1790-4182-8c53-4e78f2c435b4" (UID: "8fe0bc14-1790-4182-8c53-4e78f2c435b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.942523 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smc24\" (UniqueName: \"kubernetes.io/projected/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-kube-api-access-smc24\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.942755 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.942816 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.942887 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.942951 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plbgf\" (UniqueName: \"kubernetes.io/projected/8fe0bc14-1790-4182-8c53-4e78f2c435b4-kube-api-access-plbgf\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.943037 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.943073 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8fe0bc14-1790-4182-8c53-4e78f2c435b4" (UID: "8fe0bc14-1790-4182-8c53-4e78f2c435b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.943096 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.943116 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.943210 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") on node \"crc\" " Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.943364 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.969008 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8fe0bc14-1790-4182-8c53-4e78f2c435b4" (UID: "8fe0bc14-1790-4182-8c53-4e78f2c435b4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.977506 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.990763 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8fe0bc14-1790-4182-8c53-4e78f2c435b4" (UID: "8fe0bc14-1790-4182-8c53-4e78f2c435b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.992059 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 08 19:43:17 crc kubenswrapper[4706]: I1208 19:43:17.992470 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-de119cd6-b81e-445f-a9db-62090e36b028" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028") on node "crc" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.003484 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:18 crc kubenswrapper[4706]: E1208 19:43:18.004074 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerName="glance-httpd" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004099 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerName="glance-httpd" Dec 08 19:43:18 crc kubenswrapper[4706]: E1208 19:43:18.004119 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerName="cinder-api-log" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004128 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerName="cinder-api-log" Dec 08 19:43:18 crc kubenswrapper[4706]: E1208 19:43:18.004154 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerName="glance-httpd" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004162 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerName="glance-httpd" Dec 08 19:43:18 crc kubenswrapper[4706]: E1208 19:43:18.004171 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerName="glance-log" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004178 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerName="glance-log" Dec 08 19:43:18 crc kubenswrapper[4706]: E1208 19:43:18.004193 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerName="cinder-api" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004200 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerName="cinder-api" Dec 08 19:43:18 crc kubenswrapper[4706]: E1208 19:43:18.004207 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" containerName="dnsmasq-dns" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004214 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" containerName="dnsmasq-dns" Dec 08 19:43:18 crc kubenswrapper[4706]: E1208 19:43:18.004230 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerName="glance-log" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004237 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerName="glance-log" Dec 08 19:43:18 crc kubenswrapper[4706]: E1208 19:43:18.004254 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" containerName="init" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004278 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" containerName="init" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004474 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerName="glance-httpd" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004490 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerName="cinder-api-log" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004503 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" containerName="cinder-api" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004518 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerName="glance-log" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004533 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" containerName="dnsmasq-dns" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004542 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" containerName="glance-httpd" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.004550 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" containerName="glance-log" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.005920 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.016459 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.016467 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.041527 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-config-data" (OuterVolumeSpecName: "config-data") pod "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" (UID: "f7fe0ac0-5907-49a9-89bb-ee12bdc8961c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.046573 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.046628 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.046654 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.046674 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.046695 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fe0bc14-1790-4182-8c53-4e78f2c435b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.046645 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.065683 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.088335 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.119393 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-bd49n"] Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.149023 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.149600 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-logs\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.149650 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.149720 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkg78\" (UniqueName: \"kubernetes.io/projected/e565173b-b249-4a74-b11e-92393ba53ac3-kube-api-access-bkg78\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.149770 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.149805 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.149830 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.149858 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.252229 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.252366 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.252397 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.252428 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.252489 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.252586 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-logs\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.252643 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.252709 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkg78\" (UniqueName: \"kubernetes.io/projected/e565173b-b249-4a74-b11e-92393ba53ac3-kube-api-access-bkg78\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.253150 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.255685 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-logs\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.257958 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.258032 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b040d665527e04c44fa620444bb13c4fde426c5d08f49faa6c04df54fad7a449/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.258517 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.266231 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.267599 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.268541 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.285297 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkg78\" (UniqueName: \"kubernetes.io/projected/e565173b-b249-4a74-b11e-92393ba53ac3-kube-api-access-bkg78\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.344521 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.596179 4706 generic.go:334] "Generic (PLEG): container finished" podID="0fe2563a-a800-40b6-84e6-60249516049f" containerID="8206afee5c336eaf2932dd8e51d32478ab5309323c56c29a56d09336c13fb877" exitCode=0 Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.596511 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" event={"ID":"0fe2563a-a800-40b6-84e6-60249516049f","Type":"ContainerDied","Data":"8206afee5c336eaf2932dd8e51d32478ab5309323c56c29a56d09336c13fb877"} Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.596990 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" event={"ID":"0fe2563a-a800-40b6-84e6-60249516049f","Type":"ContainerStarted","Data":"d406d74d6b593282ecb796de82b926be9fbc988a4769fea75ed49644fba25c4e"} Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.603336 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"f2cf99b6-8613-4e30-97f4-99317f5780d8","Type":"ContainerStarted","Data":"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734"} Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.603395 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"f2cf99b6-8613-4e30-97f4-99317f5780d8","Type":"ContainerStarted","Data":"e5dfb130ee805dfafe06dac47345b2c36a14bce25d41f693f812959fbd476aa1"} Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.614753 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" event={"ID":"8fe0bc14-1790-4182-8c53-4e78f2c435b4","Type":"ContainerDied","Data":"3a0fe7298466080f637c0c8654e6576c8b2218d978f5195345082a2d04e99786"} Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.614834 4706 scope.go:117] "RemoveContainer" containerID="308973cbb8adda8e5cbccd2e780ac364c5aa1e4179a6c79616b71a25f3d27c80" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.615186 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-t2s6s" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.640069 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7fe0ac0-5907-49a9-89bb-ee12bdc8961c","Type":"ContainerDied","Data":"d47701fd0c7dfe4a7a55eb956dc8d839d6d14fa4ffed74d0884074c444cbf352"} Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.641086 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.644664 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"4613f56e-dca1-459e-8a62-f09f52a53eb8","Type":"ContainerStarted","Data":"4f4fc07f20a632cd03f8aa9622bf11b0165457497fe14b73b22d2fa34da20aa4"} Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.648885 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.826412 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-t2s6s"] Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.838467 4706 scope.go:117] "RemoveContainer" containerID="308b1174f4d0dcacafe41dee6beebc2e3e8af97ca5387600777d14db5e00e238" Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.860625 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-t2s6s"] Dec 08 19:43:18 crc kubenswrapper[4706]: I1208 19:43:18.992356 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:18.996765 4706 scope.go:117] "RemoveContainer" containerID="4a415fe4709f1fa7a78e614f3b8db4232e11c0f2ebb2390d0d9d350dab6bc86f" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.043349 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.072337 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.078188 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.083625 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.083970 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.098780 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.127460 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.147409 4706 scope.go:117] "RemoveContainer" containerID="7ba2d2cc043b5e2d0aba651d7b6f4524658bde9741005f5bf327fe502ec52595" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.208600 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-logs\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.208675 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.209316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.209598 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.209852 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.209880 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.209915 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd2kg\" (UniqueName: \"kubernetes.io/projected/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-kube-api-access-gd2kg\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.210721 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.312960 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.313042 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.313107 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.313137 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.313158 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd2kg\" (UniqueName: \"kubernetes.io/projected/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-kube-api-access-gd2kg\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.313223 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.313248 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-logs\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.313295 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.313937 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.322675 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-logs\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.327524 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.338584 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.339566 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.340067 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.365178 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd2kg\" (UniqueName: \"kubernetes.io/projected/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-kube-api-access-gd2kg\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.476595 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.476660 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/67f02492c48f5b6b986fa4270f8226cd0203948331bc87047321fea71b95adc3/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.637466 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6144d9b7-058e-47f2-b1f8-591554ec326c" path="/var/lib/kubelet/pods/6144d9b7-058e-47f2-b1f8-591554ec326c/volumes" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.639105 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe0bc14-1790-4182-8c53-4e78f2c435b4" path="/var/lib/kubelet/pods/8fe0bc14-1790-4182-8c53-4e78f2c435b4/volumes" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.639966 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7fe0ac0-5907-49a9-89bb-ee12bdc8961c" path="/var/lib/kubelet/pods/f7fe0ac0-5907-49a9-89bb-ee12bdc8961c/volumes" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.702839 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerStarted","Data":"31561f6236a83ea3ce86f686793f5c03fc8e1f156bc9c90c8f6f653d8b6b086b"} Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.704512 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.726540 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" event={"ID":"0fe2563a-a800-40b6-84e6-60249516049f","Type":"ContainerStarted","Data":"991133f9b78a74b75ac109af725526c67db3191cec10f6be63ceedc3891720fb"} Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.739940 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.760060 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"f2cf99b6-8613-4e30-97f4-99317f5780d8","Type":"ContainerStarted","Data":"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307"} Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.760162 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.768278 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.790396 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.208522733 podStartE2EDuration="7.790372918s" podCreationTimestamp="2025-12-08 19:43:12 +0000 UTC" firstStartedPulling="2025-12-08 19:43:13.958601528 +0000 UTC m=+1276.600802531" lastFinishedPulling="2025-12-08 19:43:18.540451713 +0000 UTC m=+1281.182652716" observedRunningTime="2025-12-08 19:43:19.785437057 +0000 UTC m=+1282.427638050" watchObservedRunningTime="2025-12-08 19:43:19.790372918 +0000 UTC m=+1282.432573921" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.813101 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " pod="openstack/glance-default-external-api-0" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.840252 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" podStartSLOduration=4.840214326 podStartE2EDuration="4.840214326s" podCreationTimestamp="2025-12-08 19:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:19.831832096 +0000 UTC m=+1282.474033099" watchObservedRunningTime="2025-12-08 19:43:19.840214326 +0000 UTC m=+1282.482415329" Dec 08 19:43:19 crc kubenswrapper[4706]: I1208 19:43:19.879545 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=4.8795204210000005 podStartE2EDuration="4.879520421s" podCreationTimestamp="2025-12-08 19:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:19.856494602 +0000 UTC m=+1282.498695605" watchObservedRunningTime="2025-12-08 19:43:19.879520421 +0000 UTC m=+1282.521721414" Dec 08 19:43:20 crc kubenswrapper[4706]: I1208 19:43:20.030112 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:43:20 crc kubenswrapper[4706]: I1208 19:43:20.778816 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"4613f56e-dca1-459e-8a62-f09f52a53eb8","Type":"ContainerStarted","Data":"b02b5e8a96fc984ebdaef9b9e7ac5e331921df6e19991d57f17b88dbbee05229"} Dec 08 19:43:20 crc kubenswrapper[4706]: I1208 19:43:20.791376 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e565173b-b249-4a74-b11e-92393ba53ac3","Type":"ContainerStarted","Data":"721e7d401fd2d51024d9552f82bb5a23620971cd18aee143c874a5c0c361d3b0"} Dec 08 19:43:20 crc kubenswrapper[4706]: I1208 19:43:20.791726 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerName="cloudkitty-api-log" containerID="cri-o://65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734" gracePeriod=30 Dec 08 19:43:20 crc kubenswrapper[4706]: I1208 19:43:20.791946 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerName="cloudkitty-api" containerID="cri-o://05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307" gracePeriod=30 Dec 08 19:43:20 crc kubenswrapper[4706]: I1208 19:43:20.831508 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=3.664612449 podStartE2EDuration="5.831479004s" podCreationTimestamp="2025-12-08 19:43:15 +0000 UTC" firstStartedPulling="2025-12-08 19:43:18.120381033 +0000 UTC m=+1280.762582036" lastFinishedPulling="2025-12-08 19:43:20.287247588 +0000 UTC m=+1282.929448591" observedRunningTime="2025-12-08 19:43:20.812657565 +0000 UTC m=+1283.454858588" watchObservedRunningTime="2025-12-08 19:43:20.831479004 +0000 UTC m=+1283.473680007" Dec 08 19:43:20 crc kubenswrapper[4706]: I1208 19:43:20.843290 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:20 crc kubenswrapper[4706]: I1208 19:43:20.884712 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.774970 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.819599 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq426\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-kube-api-access-pq426\") pod \"f2cf99b6-8613-4e30-97f4-99317f5780d8\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.822797 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data\") pod \"f2cf99b6-8613-4e30-97f4-99317f5780d8\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.823024 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-certs\") pod \"f2cf99b6-8613-4e30-97f4-99317f5780d8\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.823164 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-scripts\") pod \"f2cf99b6-8613-4e30-97f4-99317f5780d8\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.823245 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data-custom\") pod \"f2cf99b6-8613-4e30-97f4-99317f5780d8\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.823396 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-combined-ca-bundle\") pod \"f2cf99b6-8613-4e30-97f4-99317f5780d8\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.823647 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cf99b6-8613-4e30-97f4-99317f5780d8-logs\") pod \"f2cf99b6-8613-4e30-97f4-99317f5780d8\" (UID: \"f2cf99b6-8613-4e30-97f4-99317f5780d8\") " Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.825205 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2cf99b6-8613-4e30-97f4-99317f5780d8-logs" (OuterVolumeSpecName: "logs") pod "f2cf99b6-8613-4e30-97f4-99317f5780d8" (UID: "f2cf99b6-8613-4e30-97f4-99317f5780d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.828562 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-kube-api-access-pq426" (OuterVolumeSpecName: "kube-api-access-pq426") pod "f2cf99b6-8613-4e30-97f4-99317f5780d8" (UID: "f2cf99b6-8613-4e30-97f4-99317f5780d8"). InnerVolumeSpecName "kube-api-access-pq426". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.830473 4706 generic.go:334] "Generic (PLEG): container finished" podID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerID="05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307" exitCode=0 Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.830590 4706 generic.go:334] "Generic (PLEG): container finished" podID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerID="65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734" exitCode=143 Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.831247 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"f2cf99b6-8613-4e30-97f4-99317f5780d8","Type":"ContainerDied","Data":"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307"} Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.831431 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"f2cf99b6-8613-4e30-97f4-99317f5780d8","Type":"ContainerDied","Data":"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734"} Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.831897 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"f2cf99b6-8613-4e30-97f4-99317f5780d8","Type":"ContainerDied","Data":"e5dfb130ee805dfafe06dac47345b2c36a14bce25d41f693f812959fbd476aa1"} Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.831973 4706 scope.go:117] "RemoveContainer" containerID="05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.833491 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.838536 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-certs" (OuterVolumeSpecName: "certs") pod "f2cf99b6-8613-4e30-97f4-99317f5780d8" (UID: "f2cf99b6-8613-4e30-97f4-99317f5780d8"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.845417 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea","Type":"ContainerStarted","Data":"7e4587a2da0934a496a2efeb6f7e7c9883137ddfe15be35abf89aa9d8d24a099"} Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.845462 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f2cf99b6-8613-4e30-97f4-99317f5780d8" (UID: "f2cf99b6-8613-4e30-97f4-99317f5780d8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.863544 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e565173b-b249-4a74-b11e-92393ba53ac3","Type":"ContainerStarted","Data":"7c7a3136ddfe2fffec794311b2f8b777789da0e5fb2cba727560ec3e903035b5"} Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.872237 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2cf99b6-8613-4e30-97f4-99317f5780d8" (UID: "f2cf99b6-8613-4e30-97f4-99317f5780d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.882221 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-scripts" (OuterVolumeSpecName: "scripts") pod "f2cf99b6-8613-4e30-97f4-99317f5780d8" (UID: "f2cf99b6-8613-4e30-97f4-99317f5780d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.927723 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.927768 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.927783 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.927800 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cf99b6-8613-4e30-97f4-99317f5780d8-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.927814 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq426\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-kube-api-access-pq426\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.927826 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f2cf99b6-8613-4e30-97f4-99317f5780d8-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:21 crc kubenswrapper[4706]: I1208 19:43:21.984640 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data" (OuterVolumeSpecName: "config-data") pod "f2cf99b6-8613-4e30-97f4-99317f5780d8" (UID: "f2cf99b6-8613-4e30-97f4-99317f5780d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.030116 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cf99b6-8613-4e30-97f4-99317f5780d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.077578 4706 scope.go:117] "RemoveContainer" containerID="65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.102449 4706 scope.go:117] "RemoveContainer" containerID="05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307" Dec 08 19:43:22 crc kubenswrapper[4706]: E1208 19:43:22.106073 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307\": container with ID starting with 05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307 not found: ID does not exist" containerID="05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.106122 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307"} err="failed to get container status \"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307\": rpc error: code = NotFound desc = could not find container \"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307\": container with ID starting with 05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307 not found: ID does not exist" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.106152 4706 scope.go:117] "RemoveContainer" containerID="65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734" Dec 08 19:43:22 crc kubenswrapper[4706]: E1208 19:43:22.106588 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734\": container with ID starting with 65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734 not found: ID does not exist" containerID="65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.106633 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734"} err="failed to get container status \"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734\": rpc error: code = NotFound desc = could not find container \"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734\": container with ID starting with 65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734 not found: ID does not exist" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.106648 4706 scope.go:117] "RemoveContainer" containerID="05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.106896 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307"} err="failed to get container status \"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307\": rpc error: code = NotFound desc = could not find container \"05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307\": container with ID starting with 05db8d4aff14e80cbc6c73acc8a3cc98f3f4b2485580f940445eadbba4035307 not found: ID does not exist" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.106914 4706 scope.go:117] "RemoveContainer" containerID="65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.107134 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734"} err="failed to get container status \"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734\": rpc error: code = NotFound desc = could not find container \"65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734\": container with ID starting with 65b1bdc2a0457f3c95614913f00ebc251d584d5b25df26f44b3f3cf0ba951734 not found: ID does not exist" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.179947 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.193920 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.248655 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:22 crc kubenswrapper[4706]: E1208 19:43:22.250794 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerName="cloudkitty-api-log" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.251214 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerName="cloudkitty-api-log" Dec 08 19:43:22 crc kubenswrapper[4706]: E1208 19:43:22.251238 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerName="cloudkitty-api" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.251249 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerName="cloudkitty-api" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.251570 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerName="cloudkitty-api-log" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.251619 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" containerName="cloudkitty-api" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.267738 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.273606 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.273869 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.274028 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.304616 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.344927 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.344998 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.345048 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.345096 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5nz8\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-kube-api-access-d5nz8\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.345120 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-scripts\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.345162 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.345182 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3863d194-54b8-4d2c-b605-4a6f19a51f49-logs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.345216 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.345277 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.447747 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.447867 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.447908 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.447953 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.447998 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5nz8\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-kube-api-access-d5nz8\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.448020 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-scripts\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.448056 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.448072 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3863d194-54b8-4d2c-b605-4a6f19a51f49-logs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.448109 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.455588 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.456021 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3863d194-54b8-4d2c-b605-4a6f19a51f49-logs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.467116 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.467339 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.467776 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.486918 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.487287 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-scripts\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.496939 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.500985 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5nz8\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-kube-api-access-d5nz8\") pod \"cloudkitty-api-0\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.617091 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.923543 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea","Type":"ContainerStarted","Data":"acbdd1c35d016d77141180892bdb89076707fb1cb7d251b680a9452396dab032"} Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.947537 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e565173b-b249-4a74-b11e-92393ba53ac3","Type":"ContainerStarted","Data":"51656795812e56dbc1e519431891b403757e847c1cc83f6cbc46b42c3550f121"} Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.954837 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="4613f56e-dca1-459e-8a62-f09f52a53eb8" containerName="cloudkitty-proc" containerID="cri-o://b02b5e8a96fc984ebdaef9b9e7ac5e331921df6e19991d57f17b88dbbee05229" gracePeriod=30 Dec 08 19:43:22 crc kubenswrapper[4706]: I1208 19:43:22.984155 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.984127121 podStartE2EDuration="5.984127121s" podCreationTimestamp="2025-12-08 19:43:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:22.978740136 +0000 UTC m=+1285.620941159" watchObservedRunningTime="2025-12-08 19:43:22.984127121 +0000 UTC m=+1285.626328124" Dec 08 19:43:23 crc kubenswrapper[4706]: I1208 19:43:23.346924 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 08 19:43:23 crc kubenswrapper[4706]: I1208 19:43:23.451003 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:23 crc kubenswrapper[4706]: I1208 19:43:23.627901 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2cf99b6-8613-4e30-97f4-99317f5780d8" path="/var/lib/kubelet/pods/f2cf99b6-8613-4e30-97f4-99317f5780d8/volumes" Dec 08 19:43:23 crc kubenswrapper[4706]: I1208 19:43:23.900481 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:43:23 crc kubenswrapper[4706]: W1208 19:43:23.900697 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3863d194_54b8_4d2c_b605_4a6f19a51f49.slice/crio-5ed014b11d0d11b2a54ffafeac05bae67971e6fe47b2cb9fe987cb5f58470f2e WatchSource:0}: Error finding container 5ed014b11d0d11b2a54ffafeac05bae67971e6fe47b2cb9fe987cb5f58470f2e: Status 404 returned error can't find the container with id 5ed014b11d0d11b2a54ffafeac05bae67971e6fe47b2cb9fe987cb5f58470f2e Dec 08 19:43:23 crc kubenswrapper[4706]: I1208 19:43:23.985931 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"3863d194-54b8-4d2c-b605-4a6f19a51f49","Type":"ContainerStarted","Data":"5ed014b11d0d11b2a54ffafeac05bae67971e6fe47b2cb9fe987cb5f58470f2e"} Dec 08 19:43:23 crc kubenswrapper[4706]: I1208 19:43:23.996466 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea","Type":"ContainerStarted","Data":"0388e89c6d2f7c592342fa82247c63861f3a445ae230f096b01075abd5742a5a"} Dec 08 19:43:23 crc kubenswrapper[4706]: I1208 19:43:23.997291 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a3505923-b491-491d-8c83-5a678f23c807" containerName="cinder-scheduler" containerID="cri-o://4dcc8bfad25c9e3360be737cbd4c1a87e35a9121184b4f42357a416193b9a192" gracePeriod=30 Dec 08 19:43:23 crc kubenswrapper[4706]: I1208 19:43:23.997522 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a3505923-b491-491d-8c83-5a678f23c807" containerName="probe" containerID="cri-o://113e4502b6e01a73f9bb145bee24b9906afb3a9a56d04bfafd2c84bda06eacf0" gracePeriod=30 Dec 08 19:43:24 crc kubenswrapper[4706]: I1208 19:43:24.050329 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.050303344 podStartE2EDuration="6.050303344s" podCreationTimestamp="2025-12-08 19:43:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:24.021733626 +0000 UTC m=+1286.663934619" watchObservedRunningTime="2025-12-08 19:43:24.050303344 +0000 UTC m=+1286.692504347" Dec 08 19:43:24 crc kubenswrapper[4706]: I1208 19:43:24.363757 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:43:24 crc kubenswrapper[4706]: I1208 19:43:24.377521 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7494d94d4d-hn8rd" Dec 08 19:43:25 crc kubenswrapper[4706]: I1208 19:43:25.004054 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-786cfc44fd-zdf7t" Dec 08 19:43:25 crc kubenswrapper[4706]: I1208 19:43:25.018239 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"3863d194-54b8-4d2c-b605-4a6f19a51f49","Type":"ContainerStarted","Data":"5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849"} Dec 08 19:43:25 crc kubenswrapper[4706]: I1208 19:43:25.018343 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 08 19:43:25 crc kubenswrapper[4706]: I1208 19:43:25.018359 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"3863d194-54b8-4d2c-b605-4a6f19a51f49","Type":"ContainerStarted","Data":"d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4"} Dec 08 19:43:25 crc kubenswrapper[4706]: I1208 19:43:25.087761 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=3.087732123 podStartE2EDuration="3.087732123s" podCreationTimestamp="2025-12-08 19:43:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:25.08030078 +0000 UTC m=+1287.722501783" watchObservedRunningTime="2025-12-08 19:43:25.087732123 +0000 UTC m=+1287.729933126" Dec 08 19:43:26 crc kubenswrapper[4706]: I1208 19:43:26.028819 4706 generic.go:334] "Generic (PLEG): container finished" podID="a3505923-b491-491d-8c83-5a678f23c807" containerID="113e4502b6e01a73f9bb145bee24b9906afb3a9a56d04bfafd2c84bda06eacf0" exitCode=0 Dec 08 19:43:26 crc kubenswrapper[4706]: I1208 19:43:26.028903 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3505923-b491-491d-8c83-5a678f23c807","Type":"ContainerDied","Data":"113e4502b6e01a73f9bb145bee24b9906afb3a9a56d04bfafd2c84bda06eacf0"} Dec 08 19:43:26 crc kubenswrapper[4706]: I1208 19:43:26.456590 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:43:26 crc kubenswrapper[4706]: I1208 19:43:26.545152 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff8449c8c-q6hd4"] Dec 08 19:43:26 crc kubenswrapper[4706]: I1208 19:43:26.545569 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" podUID="67e65750-2f38-473b-9f5e-ea2edff83e6d" containerName="dnsmasq-dns" containerID="cri-o://9f7a8791f2990524985fb50d843433f9892dc637664d1b9bc0242b2fedf6bfc8" gracePeriod=10 Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.063644 4706 generic.go:334] "Generic (PLEG): container finished" podID="67e65750-2f38-473b-9f5e-ea2edff83e6d" containerID="9f7a8791f2990524985fb50d843433f9892dc637664d1b9bc0242b2fedf6bfc8" exitCode=0 Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.063854 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" event={"ID":"67e65750-2f38-473b-9f5e-ea2edff83e6d","Type":"ContainerDied","Data":"9f7a8791f2990524985fb50d843433f9892dc637664d1b9bc0242b2fedf6bfc8"} Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.072120 4706 generic.go:334] "Generic (PLEG): container finished" podID="a3505923-b491-491d-8c83-5a678f23c807" containerID="4dcc8bfad25c9e3360be737cbd4c1a87e35a9121184b4f42357a416193b9a192" exitCode=0 Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.072190 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3505923-b491-491d-8c83-5a678f23c807","Type":"ContainerDied","Data":"4dcc8bfad25c9e3360be737cbd4c1a87e35a9121184b4f42357a416193b9a192"} Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.359533 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.386560 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-svc\") pod \"67e65750-2f38-473b-9f5e-ea2edff83e6d\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.386627 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq56b\" (UniqueName: \"kubernetes.io/projected/67e65750-2f38-473b-9f5e-ea2edff83e6d-kube-api-access-vq56b\") pod \"67e65750-2f38-473b-9f5e-ea2edff83e6d\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.386688 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-nb\") pod \"67e65750-2f38-473b-9f5e-ea2edff83e6d\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.386807 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-sb\") pod \"67e65750-2f38-473b-9f5e-ea2edff83e6d\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.386856 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-config\") pod \"67e65750-2f38-473b-9f5e-ea2edff83e6d\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.386973 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-swift-storage-0\") pod \"67e65750-2f38-473b-9f5e-ea2edff83e6d\" (UID: \"67e65750-2f38-473b-9f5e-ea2edff83e6d\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.404959 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e65750-2f38-473b-9f5e-ea2edff83e6d-kube-api-access-vq56b" (OuterVolumeSpecName: "kube-api-access-vq56b") pod "67e65750-2f38-473b-9f5e-ea2edff83e6d" (UID: "67e65750-2f38-473b-9f5e-ea2edff83e6d"). InnerVolumeSpecName "kube-api-access-vq56b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.473650 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.505652 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67e65750-2f38-473b-9f5e-ea2edff83e6d" (UID: "67e65750-2f38-473b-9f5e-ea2edff83e6d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.505847 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.505881 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq56b\" (UniqueName: \"kubernetes.io/projected/67e65750-2f38-473b-9f5e-ea2edff83e6d-kube-api-access-vq56b\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.518978 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "67e65750-2f38-473b-9f5e-ea2edff83e6d" (UID: "67e65750-2f38-473b-9f5e-ea2edff83e6d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.542396 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "67e65750-2f38-473b-9f5e-ea2edff83e6d" (UID: "67e65750-2f38-473b-9f5e-ea2edff83e6d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.557469 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "67e65750-2f38-473b-9f5e-ea2edff83e6d" (UID: "67e65750-2f38-473b-9f5e-ea2edff83e6d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.598252 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-config" (OuterVolumeSpecName: "config") pod "67e65750-2f38-473b-9f5e-ea2edff83e6d" (UID: "67e65750-2f38-473b-9f5e-ea2edff83e6d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.607794 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data-custom\") pod \"a3505923-b491-491d-8c83-5a678f23c807\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.607863 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-combined-ca-bundle\") pod \"a3505923-b491-491d-8c83-5a678f23c807\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.607964 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp5bm\" (UniqueName: \"kubernetes.io/projected/a3505923-b491-491d-8c83-5a678f23c807-kube-api-access-tp5bm\") pod \"a3505923-b491-491d-8c83-5a678f23c807\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.608015 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data\") pod \"a3505923-b491-491d-8c83-5a678f23c807\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.608049 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3505923-b491-491d-8c83-5a678f23c807-etc-machine-id\") pod \"a3505923-b491-491d-8c83-5a678f23c807\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.608138 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-scripts\") pod \"a3505923-b491-491d-8c83-5a678f23c807\" (UID: \"a3505923-b491-491d-8c83-5a678f23c807\") " Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.608886 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.608919 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.608936 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.608950 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67e65750-2f38-473b-9f5e-ea2edff83e6d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.612570 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3505923-b491-491d-8c83-5a678f23c807-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a3505923-b491-491d-8c83-5a678f23c807" (UID: "a3505923-b491-491d-8c83-5a678f23c807"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.614743 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-scripts" (OuterVolumeSpecName: "scripts") pod "a3505923-b491-491d-8c83-5a678f23c807" (UID: "a3505923-b491-491d-8c83-5a678f23c807"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.614798 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a3505923-b491-491d-8c83-5a678f23c807" (UID: "a3505923-b491-491d-8c83-5a678f23c807"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.618451 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3505923-b491-491d-8c83-5a678f23c807-kube-api-access-tp5bm" (OuterVolumeSpecName: "kube-api-access-tp5bm") pod "a3505923-b491-491d-8c83-5a678f23c807" (UID: "a3505923-b491-491d-8c83-5a678f23c807"). InnerVolumeSpecName "kube-api-access-tp5bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.684620 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3505923-b491-491d-8c83-5a678f23c807" (UID: "a3505923-b491-491d-8c83-5a678f23c807"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.711080 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.711141 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.711154 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp5bm\" (UniqueName: \"kubernetes.io/projected/a3505923-b491-491d-8c83-5a678f23c807-kube-api-access-tp5bm\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.711169 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3505923-b491-491d-8c83-5a678f23c807-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.711180 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.764480 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data" (OuterVolumeSpecName: "config-data") pod "a3505923-b491-491d-8c83-5a678f23c807" (UID: "a3505923-b491-491d-8c83-5a678f23c807"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:27 crc kubenswrapper[4706]: I1208 19:43:27.813863 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3505923-b491-491d-8c83-5a678f23c807-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.094363 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3505923-b491-491d-8c83-5a678f23c807","Type":"ContainerDied","Data":"ddec2e7ee9bf9f2e6f7e542b5647395af42bbf3c94308d43dc25775cc203e381"} Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.094446 4706 scope.go:117] "RemoveContainer" containerID="113e4502b6e01a73f9bb145bee24b9906afb3a9a56d04bfafd2c84bda06eacf0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.094429 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.097453 4706 generic.go:334] "Generic (PLEG): container finished" podID="4613f56e-dca1-459e-8a62-f09f52a53eb8" containerID="b02b5e8a96fc984ebdaef9b9e7ac5e331921df6e19991d57f17b88dbbee05229" exitCode=0 Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.097534 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"4613f56e-dca1-459e-8a62-f09f52a53eb8","Type":"ContainerDied","Data":"b02b5e8a96fc984ebdaef9b9e7ac5e331921df6e19991d57f17b88dbbee05229"} Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.102474 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" event={"ID":"67e65750-2f38-473b-9f5e-ea2edff83e6d","Type":"ContainerDied","Data":"6cdcccd964f61889725c35a419e387711343358ffcd1cf6be14d607a07a852e5"} Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.102642 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff8449c8c-q6hd4" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.136591 4706 scope.go:117] "RemoveContainer" containerID="4dcc8bfad25c9e3360be737cbd4c1a87e35a9121184b4f42357a416193b9a192" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.144810 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff8449c8c-q6hd4"] Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.161254 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ff8449c8c-q6hd4"] Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.177727 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.196963 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.226137 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:28 crc kubenswrapper[4706]: E1208 19:43:28.226931 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3505923-b491-491d-8c83-5a678f23c807" containerName="cinder-scheduler" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.226953 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3505923-b491-491d-8c83-5a678f23c807" containerName="cinder-scheduler" Dec 08 19:43:28 crc kubenswrapper[4706]: E1208 19:43:28.226965 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e65750-2f38-473b-9f5e-ea2edff83e6d" containerName="dnsmasq-dns" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.226974 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e65750-2f38-473b-9f5e-ea2edff83e6d" containerName="dnsmasq-dns" Dec 08 19:43:28 crc kubenswrapper[4706]: E1208 19:43:28.226996 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e65750-2f38-473b-9f5e-ea2edff83e6d" containerName="init" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.227004 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e65750-2f38-473b-9f5e-ea2edff83e6d" containerName="init" Dec 08 19:43:28 crc kubenswrapper[4706]: E1208 19:43:28.227016 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3505923-b491-491d-8c83-5a678f23c807" containerName="probe" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.227024 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3505923-b491-491d-8c83-5a678f23c807" containerName="probe" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.227329 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3505923-b491-491d-8c83-5a678f23c807" containerName="probe" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.227346 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3505923-b491-491d-8c83-5a678f23c807" containerName="cinder-scheduler" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.227364 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e65750-2f38-473b-9f5e-ea2edff83e6d" containerName="dnsmasq-dns" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.231557 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.236051 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.237630 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.237960 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.238446 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-z4r6h" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.241541 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.245815 4706 scope.go:117] "RemoveContainer" containerID="9f7a8791f2990524985fb50d843433f9892dc637664d1b9bc0242b2fedf6bfc8" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.334553 4706 scope.go:117] "RemoveContainer" containerID="641ad2cc7c9eb5bf5fbf725b847b3a6c7b280039acc30c0fd3bcdfc257c1093f" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.338862 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-scripts\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.338988 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.339094 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.339160 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-config-data\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.339224 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8wq6\" (UniqueName: \"kubernetes.io/projected/33d60263-f73a-470d-b0de-60eb2a8ecfd5-kube-api-access-t8wq6\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.339307 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33d60263-f73a-470d-b0de-60eb2a8ecfd5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.441636 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8wq6\" (UniqueName: \"kubernetes.io/projected/33d60263-f73a-470d-b0de-60eb2a8ecfd5-kube-api-access-t8wq6\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.441719 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33d60263-f73a-470d-b0de-60eb2a8ecfd5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.441807 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-scripts\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.441880 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.441962 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.442001 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-config-data\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.444549 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33d60263-f73a-470d-b0de-60eb2a8ecfd5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.454348 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-scripts\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.455886 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.457358 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.465565 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d60263-f73a-470d-b0de-60eb2a8ecfd5-config-data\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.472609 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8wq6\" (UniqueName: \"kubernetes.io/projected/33d60263-f73a-470d-b0de-60eb2a8ecfd5-kube-api-access-t8wq6\") pod \"cinder-scheduler-0\" (UID: \"33d60263-f73a-470d-b0de-60eb2a8ecfd5\") " pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.532476 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.543560 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrp9r\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-kube-api-access-xrp9r\") pod \"4613f56e-dca1-459e-8a62-f09f52a53eb8\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.543616 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data-custom\") pod \"4613f56e-dca1-459e-8a62-f09f52a53eb8\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.543811 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-combined-ca-bundle\") pod \"4613f56e-dca1-459e-8a62-f09f52a53eb8\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.543840 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data\") pod \"4613f56e-dca1-459e-8a62-f09f52a53eb8\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.543927 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-scripts\") pod \"4613f56e-dca1-459e-8a62-f09f52a53eb8\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.544007 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-certs\") pod \"4613f56e-dca1-459e-8a62-f09f52a53eb8\" (UID: \"4613f56e-dca1-459e-8a62-f09f52a53eb8\") " Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.549329 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-certs" (OuterVolumeSpecName: "certs") pod "4613f56e-dca1-459e-8a62-f09f52a53eb8" (UID: "4613f56e-dca1-459e-8a62-f09f52a53eb8"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.550063 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-kube-api-access-xrp9r" (OuterVolumeSpecName: "kube-api-access-xrp9r") pod "4613f56e-dca1-459e-8a62-f09f52a53eb8" (UID: "4613f56e-dca1-459e-8a62-f09f52a53eb8"). InnerVolumeSpecName "kube-api-access-xrp9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.551164 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-scripts" (OuterVolumeSpecName: "scripts") pod "4613f56e-dca1-459e-8a62-f09f52a53eb8" (UID: "4613f56e-dca1-459e-8a62-f09f52a53eb8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.551544 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4613f56e-dca1-459e-8a62-f09f52a53eb8" (UID: "4613f56e-dca1-459e-8a62-f09f52a53eb8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.565817 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.603592 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data" (OuterVolumeSpecName: "config-data") pod "4613f56e-dca1-459e-8a62-f09f52a53eb8" (UID: "4613f56e-dca1-459e-8a62-f09f52a53eb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.629101 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4613f56e-dca1-459e-8a62-f09f52a53eb8" (UID: "4613f56e-dca1-459e-8a62-f09f52a53eb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.648146 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrp9r\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-kube-api-access-xrp9r\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.648197 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.648210 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.648219 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.648231 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4613f56e-dca1-459e-8a62-f09f52a53eb8-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.648240 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4613f56e-dca1-459e-8a62-f09f52a53eb8-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.650034 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.650110 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.713676 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:28 crc kubenswrapper[4706]: I1208 19:43:28.717921 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.117231 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.117240 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"4613f56e-dca1-459e-8a62-f09f52a53eb8","Type":"ContainerDied","Data":"4f4fc07f20a632cd03f8aa9622bf11b0165457497fe14b73b22d2fa34da20aa4"} Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.117367 4706 scope.go:117] "RemoveContainer" containerID="b02b5e8a96fc984ebdaef9b9e7ac5e331921df6e19991d57f17b88dbbee05229" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.123131 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.123169 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.177777 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.191798 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.214070 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.241317 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:29 crc kubenswrapper[4706]: E1208 19:43:29.241927 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4613f56e-dca1-459e-8a62-f09f52a53eb8" containerName="cloudkitty-proc" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.241951 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4613f56e-dca1-459e-8a62-f09f52a53eb8" containerName="cloudkitty-proc" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.242278 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4613f56e-dca1-459e-8a62-f09f52a53eb8" containerName="cloudkitty-proc" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.243483 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.260858 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.261816 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.374205 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlhdc\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-kube-api-access-wlhdc\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.374308 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.374339 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.374399 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-certs\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.374464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.374509 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.476867 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.476954 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlhdc\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-kube-api-access-wlhdc\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.477012 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.477040 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.477116 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-certs\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.477194 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.488594 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.490146 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-certs\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.491506 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.499218 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.499419 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.503176 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlhdc\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-kube-api-access-wlhdc\") pod \"cloudkitty-proc-0\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.674526 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4613f56e-dca1-459e-8a62-f09f52a53eb8" path="/var/lib/kubelet/pods/4613f56e-dca1-459e-8a62-f09f52a53eb8/volumes" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.679198 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e65750-2f38-473b-9f5e-ea2edff83e6d" path="/var/lib/kubelet/pods/67e65750-2f38-473b-9f5e-ea2edff83e6d/volumes" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.680279 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3505923-b491-491d-8c83-5a678f23c807" path="/var/lib/kubelet/pods/a3505923-b491-491d-8c83-5a678f23c807/volumes" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.750046 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.848448 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.851153 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.856603 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.856915 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-4rdf5" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.857087 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.861284 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.998837 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-openstack-config-secret\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.999486 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv5m5\" (UniqueName: \"kubernetes.io/projected/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-kube-api-access-xv5m5\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.999535 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:29 crc kubenswrapper[4706]: I1208 19:43:29.999566 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-openstack-config\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.030575 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.031033 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.101755 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv5m5\" (UniqueName: \"kubernetes.io/projected/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-kube-api-access-xv5m5\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.101816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.101849 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-openstack-config\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.101936 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-openstack-config-secret\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.106077 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-openstack-config\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.106283 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.113658 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.126829 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-openstack-config-secret\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.135109 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.149066 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv5m5\" (UniqueName: \"kubernetes.io/projected/1b714f3f-e93c-4a5b-ae82-1f551b163fe6-kube-api-access-xv5m5\") pod \"openstackclient\" (UID: \"1b714f3f-e93c-4a5b-ae82-1f551b163fe6\") " pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.188503 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33d60263-f73a-470d-b0de-60eb2a8ecfd5","Type":"ContainerStarted","Data":"5883c8640973d21603fec9b06674882d641cc9a3cb347c8f6123f497ca1d21de"} Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.188604 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.190225 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.215152 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.411301 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:43:30 crc kubenswrapper[4706]: W1208 19:43:30.490116 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode460c85d_5d1f_4a1c_9cfc_0163f83a2ddb.slice/crio-63b730bfd3410a162747fdaad00aa744d58eaa0c36222a7dea3b3fa467bd405a WatchSource:0}: Error finding container 63b730bfd3410a162747fdaad00aa744d58eaa0c36222a7dea3b3fa467bd405a: Status 404 returned error can't find the container with id 63b730bfd3410a162747fdaad00aa744d58eaa0c36222a7dea3b3fa467bd405a Dec 08 19:43:30 crc kubenswrapper[4706]: I1208 19:43:30.884884 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 08 19:43:31 crc kubenswrapper[4706]: I1208 19:43:31.232045 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1b714f3f-e93c-4a5b-ae82-1f551b163fe6","Type":"ContainerStarted","Data":"559a95b43b26c1bc31ff6b1d7fc63b4b61ee8e1474181ecd0a8005b61cbcde33"} Dec 08 19:43:31 crc kubenswrapper[4706]: I1208 19:43:31.234685 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33d60263-f73a-470d-b0de-60eb2a8ecfd5","Type":"ContainerStarted","Data":"ee20b23806d668bc5f38593b13e53edad6ad8da1eb0485cfea53813189978dbc"} Dec 08 19:43:31 crc kubenswrapper[4706]: I1208 19:43:31.236966 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb","Type":"ContainerStarted","Data":"0cb9a1ad5a1bb40669096ec41b53105becfc327ee89d251fa8c5f50db97e76b9"} Dec 08 19:43:31 crc kubenswrapper[4706]: I1208 19:43:31.236991 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb","Type":"ContainerStarted","Data":"63b730bfd3410a162747fdaad00aa744d58eaa0c36222a7dea3b3fa467bd405a"} Dec 08 19:43:31 crc kubenswrapper[4706]: I1208 19:43:31.237189 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:43:31 crc kubenswrapper[4706]: I1208 19:43:31.237243 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:43:31 crc kubenswrapper[4706]: I1208 19:43:31.286392 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.28635531 podStartE2EDuration="2.28635531s" podCreationTimestamp="2025-12-08 19:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:31.267946013 +0000 UTC m=+1293.910147016" watchObservedRunningTime="2025-12-08 19:43:31.28635531 +0000 UTC m=+1293.928556313" Dec 08 19:43:32 crc kubenswrapper[4706]: I1208 19:43:32.251869 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33d60263-f73a-470d-b0de-60eb2a8ecfd5","Type":"ContainerStarted","Data":"ca80a4dfdaaba61461996d7468d8de70f5ee14eb53f6808209df86268a895e72"} Dec 08 19:43:32 crc kubenswrapper[4706]: I1208 19:43:32.297552 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.297528157 podStartE2EDuration="4.297528157s" podCreationTimestamp="2025-12-08 19:43:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:32.279451439 +0000 UTC m=+1294.921652452" watchObservedRunningTime="2025-12-08 19:43:32.297528157 +0000 UTC m=+1294.939729160" Dec 08 19:43:33 crc kubenswrapper[4706]: I1208 19:43:33.566548 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 08 19:43:33 crc kubenswrapper[4706]: I1208 19:43:33.713429 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:33 crc kubenswrapper[4706]: I1208 19:43:33.713595 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:43:33 crc kubenswrapper[4706]: I1208 19:43:33.757481 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 08 19:43:33 crc kubenswrapper[4706]: I1208 19:43:33.757631 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:43:33 crc kubenswrapper[4706]: I1208 19:43:33.979049 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.111493 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.777028 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6d86ddd67f-5b6vx"] Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.779662 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.784088 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.784314 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.784484 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.815422 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6d86ddd67f-5b6vx"] Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.940123 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/637d7d72-7ba8-4a35-b37c-35aef7813c75-etc-swift\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.940548 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8s7d\" (UniqueName: \"kubernetes.io/projected/637d7d72-7ba8-4a35-b37c-35aef7813c75-kube-api-access-p8s7d\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.940675 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-public-tls-certs\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.940701 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-internal-tls-certs\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.940798 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/637d7d72-7ba8-4a35-b37c-35aef7813c75-log-httpd\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.940837 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/637d7d72-7ba8-4a35-b37c-35aef7813c75-run-httpd\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.940856 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-combined-ca-bundle\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:34 crc kubenswrapper[4706]: I1208 19:43:34.940931 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-config-data\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.043764 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/637d7d72-7ba8-4a35-b37c-35aef7813c75-etc-swift\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.043849 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8s7d\" (UniqueName: \"kubernetes.io/projected/637d7d72-7ba8-4a35-b37c-35aef7813c75-kube-api-access-p8s7d\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.043942 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-public-tls-certs\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.043967 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-internal-tls-certs\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.044020 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/637d7d72-7ba8-4a35-b37c-35aef7813c75-log-httpd\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.044057 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/637d7d72-7ba8-4a35-b37c-35aef7813c75-run-httpd\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.044081 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-combined-ca-bundle\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.044157 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-config-data\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.045534 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/637d7d72-7ba8-4a35-b37c-35aef7813c75-log-httpd\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.046130 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/637d7d72-7ba8-4a35-b37c-35aef7813c75-run-httpd\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.052496 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-config-data\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.054663 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/637d7d72-7ba8-4a35-b37c-35aef7813c75-etc-swift\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.058898 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-public-tls-certs\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.059373 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-combined-ca-bundle\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.067654 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/637d7d72-7ba8-4a35-b37c-35aef7813c75-internal-tls-certs\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.067843 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8s7d\" (UniqueName: \"kubernetes.io/projected/637d7d72-7ba8-4a35-b37c-35aef7813c75-kube-api-access-p8s7d\") pod \"swift-proxy-6d86ddd67f-5b6vx\" (UID: \"637d7d72-7ba8-4a35-b37c-35aef7813c75\") " pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.147295 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:35 crc kubenswrapper[4706]: I1208 19:43:35.828339 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6d86ddd67f-5b6vx"] Dec 08 19:43:36 crc kubenswrapper[4706]: I1208 19:43:36.341213 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" event={"ID":"637d7d72-7ba8-4a35-b37c-35aef7813c75","Type":"ContainerStarted","Data":"6142ec573ea793ad0ca517c590017540b51dbaaa011528d3846c206b1b779f2e"} Dec 08 19:43:36 crc kubenswrapper[4706]: I1208 19:43:36.342151 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" event={"ID":"637d7d72-7ba8-4a35-b37c-35aef7813c75","Type":"ContainerStarted","Data":"f9932435982a0fb07edff90a8502568f67328915fd049625640d8539815c3334"} Dec 08 19:43:36 crc kubenswrapper[4706]: I1208 19:43:36.828332 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:36 crc kubenswrapper[4706]: I1208 19:43:36.828678 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="ceilometer-central-agent" containerID="cri-o://db5da6a4db8a9b74a63d29235c313b87049a39af1a933e120a14a75ccc453a05" gracePeriod=30 Dec 08 19:43:36 crc kubenswrapper[4706]: I1208 19:43:36.828765 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="proxy-httpd" containerID="cri-o://31561f6236a83ea3ce86f686793f5c03fc8e1f156bc9c90c8f6f653d8b6b086b" gracePeriod=30 Dec 08 19:43:36 crc kubenswrapper[4706]: I1208 19:43:36.828848 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="ceilometer-notification-agent" containerID="cri-o://d40db7467f3e1069452c5f73a15afa51bfda2da03ce4562dd1e2daa5a4b1c646" gracePeriod=30 Dec 08 19:43:36 crc kubenswrapper[4706]: I1208 19:43:36.828765 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="sg-core" containerID="cri-o://767012e581f3c74aeb46a9125c724ba78c8da18be184f6f10983a171033b3e12" gracePeriod=30 Dec 08 19:43:36 crc kubenswrapper[4706]: I1208 19:43:36.851618 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.180:3000/\": EOF" Dec 08 19:43:37 crc kubenswrapper[4706]: I1208 19:43:37.370137 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" event={"ID":"637d7d72-7ba8-4a35-b37c-35aef7813c75","Type":"ContainerStarted","Data":"74e527bcafeb73d6a327be9720006878c59f1baa9c41c80719807adb1c8e28b1"} Dec 08 19:43:37 crc kubenswrapper[4706]: I1208 19:43:37.370871 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:37 crc kubenswrapper[4706]: I1208 19:43:37.370904 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:37 crc kubenswrapper[4706]: I1208 19:43:37.379292 4706 generic.go:334] "Generic (PLEG): container finished" podID="f9f63de4-e7de-4a25-8044-0c2603442785" containerID="31561f6236a83ea3ce86f686793f5c03fc8e1f156bc9c90c8f6f653d8b6b086b" exitCode=0 Dec 08 19:43:37 crc kubenswrapper[4706]: I1208 19:43:37.379332 4706 generic.go:334] "Generic (PLEG): container finished" podID="f9f63de4-e7de-4a25-8044-0c2603442785" containerID="767012e581f3c74aeb46a9125c724ba78c8da18be184f6f10983a171033b3e12" exitCode=2 Dec 08 19:43:37 crc kubenswrapper[4706]: I1208 19:43:37.379366 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerDied","Data":"31561f6236a83ea3ce86f686793f5c03fc8e1f156bc9c90c8f6f653d8b6b086b"} Dec 08 19:43:37 crc kubenswrapper[4706]: I1208 19:43:37.379399 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerDied","Data":"767012e581f3c74aeb46a9125c724ba78c8da18be184f6f10983a171033b3e12"} Dec 08 19:43:37 crc kubenswrapper[4706]: I1208 19:43:37.403762 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" podStartSLOduration=3.403737979 podStartE2EDuration="3.403737979s" podCreationTimestamp="2025-12-08 19:43:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:37.393762623 +0000 UTC m=+1300.035963646" watchObservedRunningTime="2025-12-08 19:43:37.403737979 +0000 UTC m=+1300.045938982" Dec 08 19:43:38 crc kubenswrapper[4706]: I1208 19:43:38.399051 4706 generic.go:334] "Generic (PLEG): container finished" podID="f9f63de4-e7de-4a25-8044-0c2603442785" containerID="db5da6a4db8a9b74a63d29235c313b87049a39af1a933e120a14a75ccc453a05" exitCode=0 Dec 08 19:43:38 crc kubenswrapper[4706]: I1208 19:43:38.399113 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerDied","Data":"db5da6a4db8a9b74a63d29235c313b87049a39af1a933e120a14a75ccc453a05"} Dec 08 19:43:38 crc kubenswrapper[4706]: I1208 19:43:38.943156 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 08 19:43:42 crc kubenswrapper[4706]: I1208 19:43:42.485336 4706 generic.go:334] "Generic (PLEG): container finished" podID="f9f63de4-e7de-4a25-8044-0c2603442785" containerID="d40db7467f3e1069452c5f73a15afa51bfda2da03ce4562dd1e2daa5a4b1c646" exitCode=0 Dec 08 19:43:42 crc kubenswrapper[4706]: I1208 19:43:42.485369 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerDied","Data":"d40db7467f3e1069452c5f73a15afa51bfda2da03ce4562dd1e2daa5a4b1c646"} Dec 08 19:43:43 crc kubenswrapper[4706]: I1208 19:43:43.220012 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.180:3000/\": dial tcp 10.217.0.180:3000: connect: connection refused" Dec 08 19:43:45 crc kubenswrapper[4706]: I1208 19:43:45.152670 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:45 crc kubenswrapper[4706]: I1208 19:43:45.156503 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6d86ddd67f-5b6vx" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.394588 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.517762 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-config-data\") pod \"f9f63de4-e7de-4a25-8044-0c2603442785\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.517837 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-combined-ca-bundle\") pod \"f9f63de4-e7de-4a25-8044-0c2603442785\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.517999 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-sg-core-conf-yaml\") pod \"f9f63de4-e7de-4a25-8044-0c2603442785\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.518042 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-run-httpd\") pod \"f9f63de4-e7de-4a25-8044-0c2603442785\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.518068 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-scripts\") pod \"f9f63de4-e7de-4a25-8044-0c2603442785\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.518115 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-log-httpd\") pod \"f9f63de4-e7de-4a25-8044-0c2603442785\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.518159 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66ls4\" (UniqueName: \"kubernetes.io/projected/f9f63de4-e7de-4a25-8044-0c2603442785-kube-api-access-66ls4\") pod \"f9f63de4-e7de-4a25-8044-0c2603442785\" (UID: \"f9f63de4-e7de-4a25-8044-0c2603442785\") " Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.518768 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f9f63de4-e7de-4a25-8044-0c2603442785" (UID: "f9f63de4-e7de-4a25-8044-0c2603442785"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.519037 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f9f63de4-e7de-4a25-8044-0c2603442785" (UID: "f9f63de4-e7de-4a25-8044-0c2603442785"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.527573 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-scripts" (OuterVolumeSpecName: "scripts") pod "f9f63de4-e7de-4a25-8044-0c2603442785" (UID: "f9f63de4-e7de-4a25-8044-0c2603442785"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.533451 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9f63de4-e7de-4a25-8044-0c2603442785-kube-api-access-66ls4" (OuterVolumeSpecName: "kube-api-access-66ls4") pod "f9f63de4-e7de-4a25-8044-0c2603442785" (UID: "f9f63de4-e7de-4a25-8044-0c2603442785"). InnerVolumeSpecName "kube-api-access-66ls4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.556608 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f63de4-e7de-4a25-8044-0c2603442785","Type":"ContainerDied","Data":"ac6c4653f38bb585725a7a40dcf9155b061696baafcf1c22c04ecad1780c92d1"} Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.556683 4706 scope.go:117] "RemoveContainer" containerID="31561f6236a83ea3ce86f686793f5c03fc8e1f156bc9c90c8f6f653d8b6b086b" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.556883 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.565199 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1b714f3f-e93c-4a5b-ae82-1f551b163fe6","Type":"ContainerStarted","Data":"17a9429b3ad6891e9727cb91dc166f7b14ead1851ec661f0a4bfddb9a1882fd7"} Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.566555 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f9f63de4-e7de-4a25-8044-0c2603442785" (UID: "f9f63de4-e7de-4a25-8044-0c2603442785"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.593822 4706 scope.go:117] "RemoveContainer" containerID="767012e581f3c74aeb46a9125c724ba78c8da18be184f6f10983a171033b3e12" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.621986 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.604239745 podStartE2EDuration="18.621959616s" podCreationTimestamp="2025-12-08 19:43:29 +0000 UTC" firstStartedPulling="2025-12-08 19:43:30.925493335 +0000 UTC m=+1293.567694338" lastFinishedPulling="2025-12-08 19:43:46.943213216 +0000 UTC m=+1309.585414209" observedRunningTime="2025-12-08 19:43:47.620581166 +0000 UTC m=+1310.262782169" watchObservedRunningTime="2025-12-08 19:43:47.621959616 +0000 UTC m=+1310.264160619" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.624005 4706 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.624037 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.624048 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.624057 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f63de4-e7de-4a25-8044-0c2603442785-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.624065 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66ls4\" (UniqueName: \"kubernetes.io/projected/f9f63de4-e7de-4a25-8044-0c2603442785-kube-api-access-66ls4\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.673544 4706 scope.go:117] "RemoveContainer" containerID="d40db7467f3e1069452c5f73a15afa51bfda2da03ce4562dd1e2daa5a4b1c646" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.681957 4706 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod42efdc0a-1b75-4412-8e3f-a2ece0aaafe5"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod42efdc0a-1b75-4412-8e3f-a2ece0aaafe5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod42efdc0a_1b75_4412_8e3f_a2ece0aaafe5.slice" Dec 08 19:43:47 crc kubenswrapper[4706]: E1208 19:43:47.682026 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod42efdc0a-1b75-4412-8e3f-a2ece0aaafe5] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod42efdc0a-1b75-4412-8e3f-a2ece0aaafe5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod42efdc0a_1b75_4412_8e3f_a2ece0aaafe5.slice" pod="openstack/cinder-api-0" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.768591 4706 scope.go:117] "RemoveContainer" containerID="db5da6a4db8a9b74a63d29235c313b87049a39af1a933e120a14a75ccc453a05" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.768640 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9f63de4-e7de-4a25-8044-0c2603442785" (UID: "f9f63de4-e7de-4a25-8044-0c2603442785"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.829879 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.889521 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-config-data" (OuterVolumeSpecName: "config-data") pod "f9f63de4-e7de-4a25-8044-0c2603442785" (UID: "f9f63de4-e7de-4a25-8044-0c2603442785"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:43:47 crc kubenswrapper[4706]: I1208 19:43:47.932036 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f63de4-e7de-4a25-8044-0c2603442785-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.196876 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.212653 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.260460 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:48 crc kubenswrapper[4706]: E1208 19:43:48.261756 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="ceilometer-central-agent" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.261787 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="ceilometer-central-agent" Dec 08 19:43:48 crc kubenswrapper[4706]: E1208 19:43:48.261835 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="ceilometer-notification-agent" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.261846 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="ceilometer-notification-agent" Dec 08 19:43:48 crc kubenswrapper[4706]: E1208 19:43:48.261887 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="sg-core" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.261902 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="sg-core" Dec 08 19:43:48 crc kubenswrapper[4706]: E1208 19:43:48.261939 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="proxy-httpd" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.261947 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="proxy-httpd" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.282139 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="ceilometer-notification-agent" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.282487 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="ceilometer-central-agent" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.282539 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="sg-core" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.282561 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" containerName="proxy-httpd" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.288673 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.288880 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.295590 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.295881 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.362072 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.362166 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-log-httpd\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.362205 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-run-httpd\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.362306 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.362404 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhsmc\" (UniqueName: \"kubernetes.io/projected/772b326e-0ec0-47b5-948e-554ae9a0a9f8-kube-api-access-bhsmc\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.362439 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-config-data\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.362464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-scripts\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.465221 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-log-httpd\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.465355 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-run-httpd\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.465462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.465493 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhsmc\" (UniqueName: \"kubernetes.io/projected/772b326e-0ec0-47b5-948e-554ae9a0a9f8-kube-api-access-bhsmc\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.465524 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-config-data\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.465830 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-log-httpd\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.465996 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-scripts\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.466101 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.466599 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-run-httpd\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.470497 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.470617 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.471069 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-scripts\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.472623 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-config-data\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.485950 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhsmc\" (UniqueName: \"kubernetes.io/projected/772b326e-0ec0-47b5-948e-554ae9a0a9f8-kube-api-access-bhsmc\") pod \"ceilometer-0\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.581384 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.622538 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.643221 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.667106 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.685498 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.689699 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.696520 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.696777 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.697083 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.699346 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.773994 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4kdn\" (UniqueName: \"kubernetes.io/projected/81a640d0-b89a-49be-aa17-00ad9d0d444a-kube-api-access-d4kdn\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.774054 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.774087 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-scripts\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.774104 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-config-data\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.774127 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.774156 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-config-data-custom\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.774176 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81a640d0-b89a-49be-aa17-00ad9d0d444a-logs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.774193 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.774245 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81a640d0-b89a-49be-aa17-00ad9d0d444a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.878834 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81a640d0-b89a-49be-aa17-00ad9d0d444a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.879435 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4kdn\" (UniqueName: \"kubernetes.io/projected/81a640d0-b89a-49be-aa17-00ad9d0d444a-kube-api-access-d4kdn\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.879473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.879539 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-scripts\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.879616 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-config-data\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.879651 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.879681 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-config-data-custom\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.879714 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81a640d0-b89a-49be-aa17-00ad9d0d444a-logs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.879740 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.878957 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81a640d0-b89a-49be-aa17-00ad9d0d444a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.883959 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81a640d0-b89a-49be-aa17-00ad9d0d444a-logs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.887336 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.887609 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.889059 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.890702 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-config-data-custom\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.891009 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-scripts\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.901028 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81a640d0-b89a-49be-aa17-00ad9d0d444a-config-data\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:48 crc kubenswrapper[4706]: I1208 19:43:48.909899 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4kdn\" (UniqueName: \"kubernetes.io/projected/81a640d0-b89a-49be-aa17-00ad9d0d444a-kube-api-access-d4kdn\") pod \"cinder-api-0\" (UID: \"81a640d0-b89a-49be-aa17-00ad9d0d444a\") " pod="openstack/cinder-api-0" Dec 08 19:43:49 crc kubenswrapper[4706]: I1208 19:43:49.116531 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 19:43:49 crc kubenswrapper[4706]: I1208 19:43:49.253924 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:49 crc kubenswrapper[4706]: I1208 19:43:49.643325 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42efdc0a-1b75-4412-8e3f-a2ece0aaafe5" path="/var/lib/kubelet/pods/42efdc0a-1b75-4412-8e3f-a2ece0aaafe5/volumes" Dec 08 19:43:49 crc kubenswrapper[4706]: I1208 19:43:49.645624 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9f63de4-e7de-4a25-8044-0c2603442785" path="/var/lib/kubelet/pods/f9f63de4-e7de-4a25-8044-0c2603442785/volumes" Dec 08 19:43:49 crc kubenswrapper[4706]: I1208 19:43:49.646759 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerStarted","Data":"093881824f1320fc4c88f9de013d7c2c260fe2ae51eb92ed6063781f158ad220"} Dec 08 19:43:49 crc kubenswrapper[4706]: I1208 19:43:49.737642 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 08 19:43:50 crc kubenswrapper[4706]: I1208 19:43:50.627076 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerStarted","Data":"c730208833bcc9daee0a2a7cdc49b06222fa47346aab301432b7a07940541a6e"} Dec 08 19:43:50 crc kubenswrapper[4706]: I1208 19:43:50.629046 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"81a640d0-b89a-49be-aa17-00ad9d0d444a","Type":"ContainerStarted","Data":"06c4234aaffc6f53a4c5c10d7a78b617648f64ee87290ae6725d1d8e8ac35cf8"} Dec 08 19:43:51 crc kubenswrapper[4706]: I1208 19:43:51.694512 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerStarted","Data":"bdb25b3ac4dacd519fd6c9db6bdd09ba32df470c03c182bcb03941d730409bb5"} Dec 08 19:43:51 crc kubenswrapper[4706]: I1208 19:43:51.703072 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"81a640d0-b89a-49be-aa17-00ad9d0d444a","Type":"ContainerStarted","Data":"b744c432b386833effc46bfeea4c0b01cfe3695efc2b3e408d7ce09458b20684"} Dec 08 19:43:52 crc kubenswrapper[4706]: I1208 19:43:52.706814 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:43:52 crc kubenswrapper[4706]: I1208 19:43:52.720559 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerStarted","Data":"3fddb0bb8bec530329ddc573cb214b2cfe45cda067a3d5cfe8afa6436b2e551a"} Dec 08 19:43:52 crc kubenswrapper[4706]: I1208 19:43:52.724124 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"81a640d0-b89a-49be-aa17-00ad9d0d444a","Type":"ContainerStarted","Data":"b8a78fd0089ffe0cf7023c10bfd6ef90bf7c5419368ce4ab9aa583cee1b28d51"} Dec 08 19:43:52 crc kubenswrapper[4706]: I1208 19:43:52.725166 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 08 19:43:52 crc kubenswrapper[4706]: I1208 19:43:52.779811 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.77978121 podStartE2EDuration="4.77978121s" podCreationTimestamp="2025-12-08 19:43:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:43:52.756009794 +0000 UTC m=+1315.398210807" watchObservedRunningTime="2025-12-08 19:43:52.77978121 +0000 UTC m=+1315.421982203" Dec 08 19:43:53 crc kubenswrapper[4706]: I1208 19:43:53.749509 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerStarted","Data":"980cc544f2e10d8a1705551914167d4255cb8d1d9a7a087d4d438f57a54bf31c"} Dec 08 19:43:53 crc kubenswrapper[4706]: I1208 19:43:53.749993 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="ceilometer-central-agent" containerID="cri-o://c730208833bcc9daee0a2a7cdc49b06222fa47346aab301432b7a07940541a6e" gracePeriod=30 Dec 08 19:43:53 crc kubenswrapper[4706]: I1208 19:43:53.750096 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="ceilometer-notification-agent" containerID="cri-o://bdb25b3ac4dacd519fd6c9db6bdd09ba32df470c03c182bcb03941d730409bb5" gracePeriod=30 Dec 08 19:43:53 crc kubenswrapper[4706]: I1208 19:43:53.750134 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="proxy-httpd" containerID="cri-o://980cc544f2e10d8a1705551914167d4255cb8d1d9a7a087d4d438f57a54bf31c" gracePeriod=30 Dec 08 19:43:53 crc kubenswrapper[4706]: I1208 19:43:53.750148 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="sg-core" containerID="cri-o://3fddb0bb8bec530329ddc573cb214b2cfe45cda067a3d5cfe8afa6436b2e551a" gracePeriod=30 Dec 08 19:43:53 crc kubenswrapper[4706]: I1208 19:43:53.788406 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.316663367 podStartE2EDuration="5.788384673s" podCreationTimestamp="2025-12-08 19:43:48 +0000 UTC" firstStartedPulling="2025-12-08 19:43:49.285980477 +0000 UTC m=+1311.928181480" lastFinishedPulling="2025-12-08 19:43:52.757701783 +0000 UTC m=+1315.399902786" observedRunningTime="2025-12-08 19:43:53.784847571 +0000 UTC m=+1316.427048574" watchObservedRunningTime="2025-12-08 19:43:53.788384673 +0000 UTC m=+1316.430585676" Dec 08 19:43:54 crc kubenswrapper[4706]: I1208 19:43:54.764636 4706 generic.go:334] "Generic (PLEG): container finished" podID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerID="980cc544f2e10d8a1705551914167d4255cb8d1d9a7a087d4d438f57a54bf31c" exitCode=0 Dec 08 19:43:54 crc kubenswrapper[4706]: I1208 19:43:54.764709 4706 generic.go:334] "Generic (PLEG): container finished" podID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerID="3fddb0bb8bec530329ddc573cb214b2cfe45cda067a3d5cfe8afa6436b2e551a" exitCode=2 Dec 08 19:43:54 crc kubenswrapper[4706]: I1208 19:43:54.764724 4706 generic.go:334] "Generic (PLEG): container finished" podID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerID="bdb25b3ac4dacd519fd6c9db6bdd09ba32df470c03c182bcb03941d730409bb5" exitCode=0 Dec 08 19:43:54 crc kubenswrapper[4706]: I1208 19:43:54.764713 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerDied","Data":"980cc544f2e10d8a1705551914167d4255cb8d1d9a7a087d4d438f57a54bf31c"} Dec 08 19:43:54 crc kubenswrapper[4706]: I1208 19:43:54.764762 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerDied","Data":"3fddb0bb8bec530329ddc573cb214b2cfe45cda067a3d5cfe8afa6436b2e551a"} Dec 08 19:43:54 crc kubenswrapper[4706]: I1208 19:43:54.764772 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerDied","Data":"bdb25b3ac4dacd519fd6c9db6bdd09ba32df470c03c182bcb03941d730409bb5"} Dec 08 19:43:59 crc kubenswrapper[4706]: I1208 19:43:59.865758 4706 generic.go:334] "Generic (PLEG): container finished" podID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerID="c730208833bcc9daee0a2a7cdc49b06222fa47346aab301432b7a07940541a6e" exitCode=0 Dec 08 19:43:59 crc kubenswrapper[4706]: I1208 19:43:59.866500 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerDied","Data":"c730208833bcc9daee0a2a7cdc49b06222fa47346aab301432b7a07940541a6e"} Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.023515 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.149284 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhsmc\" (UniqueName: \"kubernetes.io/projected/772b326e-0ec0-47b5-948e-554ae9a0a9f8-kube-api-access-bhsmc\") pod \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.149557 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-combined-ca-bundle\") pod \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.149633 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-run-httpd\") pod \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.149859 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-config-data\") pod \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.149937 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-scripts\") pod \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.149988 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-sg-core-conf-yaml\") pod \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.150018 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-log-httpd\") pod \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\" (UID: \"772b326e-0ec0-47b5-948e-554ae9a0a9f8\") " Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.151593 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "772b326e-0ec0-47b5-948e-554ae9a0a9f8" (UID: "772b326e-0ec0-47b5-948e-554ae9a0a9f8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.152653 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "772b326e-0ec0-47b5-948e-554ae9a0a9f8" (UID: "772b326e-0ec0-47b5-948e-554ae9a0a9f8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.158575 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/772b326e-0ec0-47b5-948e-554ae9a0a9f8-kube-api-access-bhsmc" (OuterVolumeSpecName: "kube-api-access-bhsmc") pod "772b326e-0ec0-47b5-948e-554ae9a0a9f8" (UID: "772b326e-0ec0-47b5-948e-554ae9a0a9f8"). InnerVolumeSpecName "kube-api-access-bhsmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.199445 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-scripts" (OuterVolumeSpecName: "scripts") pod "772b326e-0ec0-47b5-948e-554ae9a0a9f8" (UID: "772b326e-0ec0-47b5-948e-554ae9a0a9f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.201690 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "772b326e-0ec0-47b5-948e-554ae9a0a9f8" (UID: "772b326e-0ec0-47b5-948e-554ae9a0a9f8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.253406 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.253444 4706 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.253459 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.253468 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhsmc\" (UniqueName: \"kubernetes.io/projected/772b326e-0ec0-47b5-948e-554ae9a0a9f8-kube-api-access-bhsmc\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.253478 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/772b326e-0ec0-47b5-948e-554ae9a0a9f8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.311527 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "772b326e-0ec0-47b5-948e-554ae9a0a9f8" (UID: "772b326e-0ec0-47b5-948e-554ae9a0a9f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.328414 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-config-data" (OuterVolumeSpecName: "config-data") pod "772b326e-0ec0-47b5-948e-554ae9a0a9f8" (UID: "772b326e-0ec0-47b5-948e-554ae9a0a9f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.357147 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.357191 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/772b326e-0ec0-47b5-948e-554ae9a0a9f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.884977 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"772b326e-0ec0-47b5-948e-554ae9a0a9f8","Type":"ContainerDied","Data":"093881824f1320fc4c88f9de013d7c2c260fe2ae51eb92ed6063781f158ad220"} Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.885040 4706 scope.go:117] "RemoveContainer" containerID="980cc544f2e10d8a1705551914167d4255cb8d1d9a7a087d4d438f57a54bf31c" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.885209 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:00 crc kubenswrapper[4706]: I1208 19:44:00.993281 4706 scope.go:117] "RemoveContainer" containerID="3fddb0bb8bec530329ddc573cb214b2cfe45cda067a3d5cfe8afa6436b2e551a" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.006492 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.019755 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.052357 4706 scope.go:117] "RemoveContainer" containerID="bdb25b3ac4dacd519fd6c9db6bdd09ba32df470c03c182bcb03941d730409bb5" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.052761 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:01 crc kubenswrapper[4706]: E1208 19:44:01.053348 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="ceilometer-central-agent" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.053368 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="ceilometer-central-agent" Dec 08 19:44:01 crc kubenswrapper[4706]: E1208 19:44:01.053394 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="sg-core" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.053402 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="sg-core" Dec 08 19:44:01 crc kubenswrapper[4706]: E1208 19:44:01.053415 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="ceilometer-notification-agent" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.053422 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="ceilometer-notification-agent" Dec 08 19:44:01 crc kubenswrapper[4706]: E1208 19:44:01.053436 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="proxy-httpd" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.053443 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="proxy-httpd" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.053661 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="sg-core" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.053719 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="ceilometer-central-agent" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.053735 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="ceilometer-notification-agent" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.053754 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" containerName="proxy-httpd" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.056131 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.060823 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.068628 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.096615 4706 scope.go:117] "RemoveContainer" containerID="c730208833bcc9daee0a2a7cdc49b06222fa47346aab301432b7a07940541a6e" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.110738 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.183662 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.183745 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgr9v\" (UniqueName: \"kubernetes.io/projected/8f58e55a-69a4-457c-aecf-8681f88e0ce6-kube-api-access-tgr9v\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.183826 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-config-data\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.183886 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-scripts\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.183975 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-log-httpd\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.184000 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.184028 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-run-httpd\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.320412 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-log-httpd\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.320516 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.320558 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-run-httpd\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.320678 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.320724 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgr9v\" (UniqueName: \"kubernetes.io/projected/8f58e55a-69a4-457c-aecf-8681f88e0ce6-kube-api-access-tgr9v\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.320894 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-config-data\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.321016 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-scripts\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.322880 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-log-httpd\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.325130 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-run-httpd\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.347959 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-scripts\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.350025 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.356276 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.357993 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgr9v\" (UniqueName: \"kubernetes.io/projected/8f58e55a-69a4-457c-aecf-8681f88e0ce6-kube-api-access-tgr9v\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.358503 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-config-data\") pod \"ceilometer-0\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.387003 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:01 crc kubenswrapper[4706]: I1208 19:44:01.655197 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="772b326e-0ec0-47b5-948e-554ae9a0a9f8" path="/var/lib/kubelet/pods/772b326e-0ec0-47b5-948e-554ae9a0a9f8/volumes" Dec 08 19:44:02 crc kubenswrapper[4706]: I1208 19:44:02.027833 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:02 crc kubenswrapper[4706]: I1208 19:44:02.057434 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Dec 08 19:44:02 crc kubenswrapper[4706]: I1208 19:44:02.516851 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 08 19:44:02 crc kubenswrapper[4706]: I1208 19:44:02.910673 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerStarted","Data":"57bf85b62b27f1ba5ee2aa0db4cee2def5dd29a65a5cc010e0583431398354cd"} Dec 08 19:44:02 crc kubenswrapper[4706]: I1208 19:44:02.911133 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerStarted","Data":"357460b0c13d6d5602d31d7936eea15b973b4bb29858e883f6e57f9e48495936"} Dec 08 19:44:03 crc kubenswrapper[4706]: I1208 19:44:03.967901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerStarted","Data":"5808c7b997e14db62e6f8937c59569056dee8f68330e025b20e6890b2a9772c1"} Dec 08 19:44:04 crc kubenswrapper[4706]: I1208 19:44:04.982108 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerStarted","Data":"608db59125990886f0f29367fcdd13d32c26ce2f6501e7eba2523008c9b7429c"} Dec 08 19:44:05 crc kubenswrapper[4706]: I1208 19:44:05.600447 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:44:05 crc kubenswrapper[4706]: I1208 19:44:05.608850 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerName="glance-httpd" containerID="cri-o://0388e89c6d2f7c592342fa82247c63861f3a445ae230f096b01075abd5742a5a" gracePeriod=30 Dec 08 19:44:05 crc kubenswrapper[4706]: I1208 19:44:05.609057 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerName="glance-log" containerID="cri-o://acbdd1c35d016d77141180892bdb89076707fb1cb7d251b680a9452396dab032" gracePeriod=30 Dec 08 19:44:05 crc kubenswrapper[4706]: I1208 19:44:05.996690 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerStarted","Data":"0398181274faf2faa938930a494f942266d8782f3720622e205c4e7e3ee3a8f8"} Dec 08 19:44:05 crc kubenswrapper[4706]: I1208 19:44:05.996809 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 19:44:05 crc kubenswrapper[4706]: I1208 19:44:05.999654 4706 generic.go:334] "Generic (PLEG): container finished" podID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerID="acbdd1c35d016d77141180892bdb89076707fb1cb7d251b680a9452396dab032" exitCode=143 Dec 08 19:44:05 crc kubenswrapper[4706]: I1208 19:44:05.999724 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea","Type":"ContainerDied","Data":"acbdd1c35d016d77141180892bdb89076707fb1cb7d251b680a9452396dab032"} Dec 08 19:44:06 crc kubenswrapper[4706]: I1208 19:44:06.024487 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.7373780189999999 podStartE2EDuration="5.024457839s" podCreationTimestamp="2025-12-08 19:44:01 +0000 UTC" firstStartedPulling="2025-12-08 19:44:02.030653872 +0000 UTC m=+1324.672854875" lastFinishedPulling="2025-12-08 19:44:05.317733692 +0000 UTC m=+1327.959934695" observedRunningTime="2025-12-08 19:44:06.018318861 +0000 UTC m=+1328.660519864" watchObservedRunningTime="2025-12-08 19:44:06.024457839 +0000 UTC m=+1328.666658842" Dec 08 19:44:06 crc kubenswrapper[4706]: I1208 19:44:06.105941 4706 scope.go:117] "RemoveContainer" containerID="72df8d15c56a4e99047ab67fbcf3d58f2167583cbf52f3b5f34fe992e13931d2" Dec 08 19:44:06 crc kubenswrapper[4706]: I1208 19:44:06.142956 4706 scope.go:117] "RemoveContainer" containerID="a4b86d85e243e29b2fc53f77f0bb32a4cc3ee80afc571cc6facb7fb77b4ca54c" Dec 08 19:44:06 crc kubenswrapper[4706]: I1208 19:44:06.216515 4706 scope.go:117] "RemoveContainer" containerID="9da102eddedfc5f349c5d0c4d152a9f611087a1520572d354ef4a144bc59e25a" Dec 08 19:44:09 crc kubenswrapper[4706]: I1208 19:44:09.041189 4706 generic.go:334] "Generic (PLEG): container finished" podID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerID="0388e89c6d2f7c592342fa82247c63861f3a445ae230f096b01075abd5742a5a" exitCode=0 Dec 08 19:44:09 crc kubenswrapper[4706]: I1208 19:44:09.041678 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea","Type":"ContainerDied","Data":"0388e89c6d2f7c592342fa82247c63861f3a445ae230f096b01075abd5742a5a"} Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.299428 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-kd4rj"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.302150 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.313376 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kd4rj"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.397343 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-c82mx"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.399584 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.409549 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-operator-scripts\") pod \"nova-api-db-create-kd4rj\" (UID: \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\") " pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.409600 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29bwj\" (UniqueName: \"kubernetes.io/projected/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-kube-api-access-29bwj\") pod \"nova-api-db-create-kd4rj\" (UID: \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\") " pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.432765 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-c82mx"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.490029 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-t86b4"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.493586 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.512383 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-operator-scripts\") pod \"nova-api-db-create-kd4rj\" (UID: \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\") " pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.512453 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29bwj\" (UniqueName: \"kubernetes.io/projected/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-kube-api-access-29bwj\") pod \"nova-api-db-create-kd4rj\" (UID: \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\") " pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.512504 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50786407-6513-4d04-b13b-0367252b90af-operator-scripts\") pod \"nova-cell0-db-create-c82mx\" (UID: \"50786407-6513-4d04-b13b-0367252b90af\") " pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.512607 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr85w\" (UniqueName: \"kubernetes.io/projected/50786407-6513-4d04-b13b-0367252b90af-kube-api-access-nr85w\") pod \"nova-cell0-db-create-c82mx\" (UID: \"50786407-6513-4d04-b13b-0367252b90af\") " pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.513919 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-t86b4"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.516385 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-operator-scripts\") pod \"nova-api-db-create-kd4rj\" (UID: \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\") " pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.605196 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29bwj\" (UniqueName: \"kubernetes.io/projected/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-kube-api-access-29bwj\") pod \"nova-api-db-create-kd4rj\" (UID: \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\") " pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.605304 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c56a-account-create-update-kl49q"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.607078 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.613745 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.615338 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50786407-6513-4d04-b13b-0367252b90af-operator-scripts\") pod \"nova-cell0-db-create-c82mx\" (UID: \"50786407-6513-4d04-b13b-0367252b90af\") " pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.615452 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr85w\" (UniqueName: \"kubernetes.io/projected/50786407-6513-4d04-b13b-0367252b90af-kube-api-access-nr85w\") pod \"nova-cell0-db-create-c82mx\" (UID: \"50786407-6513-4d04-b13b-0367252b90af\") " pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.615527 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv6hs\" (UniqueName: \"kubernetes.io/projected/855d2f91-5dfa-401f-b1ac-a3f94287c245-kube-api-access-sv6hs\") pod \"nova-cell1-db-create-t86b4\" (UID: \"855d2f91-5dfa-401f-b1ac-a3f94287c245\") " pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.615595 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/855d2f91-5dfa-401f-b1ac-a3f94287c245-operator-scripts\") pod \"nova-cell1-db-create-t86b4\" (UID: \"855d2f91-5dfa-401f-b1ac-a3f94287c245\") " pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.616591 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50786407-6513-4d04-b13b-0367252b90af-operator-scripts\") pod \"nova-cell0-db-create-c82mx\" (UID: \"50786407-6513-4d04-b13b-0367252b90af\") " pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.657787 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.662328 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c56a-account-create-update-kl49q"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.718857 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66571791-d210-477a-8951-20a0609193a9-operator-scripts\") pod \"nova-api-c56a-account-create-update-kl49q\" (UID: \"66571791-d210-477a-8951-20a0609193a9\") " pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.719077 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfv8b\" (UniqueName: \"kubernetes.io/projected/66571791-d210-477a-8951-20a0609193a9-kube-api-access-pfv8b\") pod \"nova-api-c56a-account-create-update-kl49q\" (UID: \"66571791-d210-477a-8951-20a0609193a9\") " pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.719313 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv6hs\" (UniqueName: \"kubernetes.io/projected/855d2f91-5dfa-401f-b1ac-a3f94287c245-kube-api-access-sv6hs\") pod \"nova-cell1-db-create-t86b4\" (UID: \"855d2f91-5dfa-401f-b1ac-a3f94287c245\") " pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.719391 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/855d2f91-5dfa-401f-b1ac-a3f94287c245-operator-scripts\") pod \"nova-cell1-db-create-t86b4\" (UID: \"855d2f91-5dfa-401f-b1ac-a3f94287c245\") " pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.720952 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/855d2f91-5dfa-401f-b1ac-a3f94287c245-operator-scripts\") pod \"nova-cell1-db-create-t86b4\" (UID: \"855d2f91-5dfa-401f-b1ac-a3f94287c245\") " pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.757167 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr85w\" (UniqueName: \"kubernetes.io/projected/50786407-6513-4d04-b13b-0367252b90af-kube-api-access-nr85w\") pod \"nova-cell0-db-create-c82mx\" (UID: \"50786407-6513-4d04-b13b-0367252b90af\") " pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.758983 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv6hs\" (UniqueName: \"kubernetes.io/projected/855d2f91-5dfa-401f-b1ac-a3f94287c245-kube-api-access-sv6hs\") pod \"nova-cell1-db-create-t86b4\" (UID: \"855d2f91-5dfa-401f-b1ac-a3f94287c245\") " pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.806336 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-758b-account-create-update-zh9x9"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.837583 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.843111 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.844947 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f7v2\" (UniqueName: \"kubernetes.io/projected/868f8ce6-c1f8-404d-8226-7fffb0518de4-kube-api-access-7f7v2\") pod \"nova-cell0-758b-account-create-update-zh9x9\" (UID: \"868f8ce6-c1f8-404d-8226-7fffb0518de4\") " pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.845061 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfv8b\" (UniqueName: \"kubernetes.io/projected/66571791-d210-477a-8951-20a0609193a9-kube-api-access-pfv8b\") pod \"nova-api-c56a-account-create-update-kl49q\" (UID: \"66571791-d210-477a-8951-20a0609193a9\") " pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.845686 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/868f8ce6-c1f8-404d-8226-7fffb0518de4-operator-scripts\") pod \"nova-cell0-758b-account-create-update-zh9x9\" (UID: \"868f8ce6-c1f8-404d-8226-7fffb0518de4\") " pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.845886 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66571791-d210-477a-8951-20a0609193a9-operator-scripts\") pod \"nova-api-c56a-account-create-update-kl49q\" (UID: \"66571791-d210-477a-8951-20a0609193a9\") " pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.847188 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66571791-d210-477a-8951-20a0609193a9-operator-scripts\") pod \"nova-api-c56a-account-create-update-kl49q\" (UID: \"66571791-d210-477a-8951-20a0609193a9\") " pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.878741 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-758b-account-create-update-zh9x9"] Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.895329 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfv8b\" (UniqueName: \"kubernetes.io/projected/66571791-d210-477a-8951-20a0609193a9-kube-api-access-pfv8b\") pod \"nova-api-c56a-account-create-update-kl49q\" (UID: \"66571791-d210-477a-8951-20a0609193a9\") " pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.897677 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.962816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f7v2\" (UniqueName: \"kubernetes.io/projected/868f8ce6-c1f8-404d-8226-7fffb0518de4-kube-api-access-7f7v2\") pod \"nova-cell0-758b-account-create-update-zh9x9\" (UID: \"868f8ce6-c1f8-404d-8226-7fffb0518de4\") " pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.963053 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/868f8ce6-c1f8-404d-8226-7fffb0518de4-operator-scripts\") pod \"nova-cell0-758b-account-create-update-zh9x9\" (UID: \"868f8ce6-c1f8-404d-8226-7fffb0518de4\") " pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:10 crc kubenswrapper[4706]: I1208 19:44:10.974241 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/868f8ce6-c1f8-404d-8226-7fffb0518de4-operator-scripts\") pod \"nova-cell0-758b-account-create-update-zh9x9\" (UID: \"868f8ce6-c1f8-404d-8226-7fffb0518de4\") " pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.018892 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f7v2\" (UniqueName: \"kubernetes.io/projected/868f8ce6-c1f8-404d-8226-7fffb0518de4-kube-api-access-7f7v2\") pod \"nova-cell0-758b-account-create-update-zh9x9\" (UID: \"868f8ce6-c1f8-404d-8226-7fffb0518de4\") " pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.039747 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.070137 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2143-account-create-update-fqwzd"] Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.070654 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.072962 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.076076 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.088535 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2143-account-create-update-fqwzd"] Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.095536 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.159556 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.191702 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-combined-ca-bundle\") pod \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.191763 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-public-tls-certs\") pod \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.191857 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd2kg\" (UniqueName: \"kubernetes.io/projected/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-kube-api-access-gd2kg\") pod \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.191948 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-httpd-run\") pod \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.192017 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-scripts\") pod \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.192041 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-config-data\") pod \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.192126 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-logs\") pod \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.192428 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\" (UID: \"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea\") " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.192851 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nkrt\" (UniqueName: \"kubernetes.io/projected/d05ed39a-048e-4de0-812c-99daca774a99-kube-api-access-8nkrt\") pod \"nova-cell1-2143-account-create-update-fqwzd\" (UID: \"d05ed39a-048e-4de0-812c-99daca774a99\") " pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.192898 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d05ed39a-048e-4de0-812c-99daca774a99-operator-scripts\") pod \"nova-cell1-2143-account-create-update-fqwzd\" (UID: \"d05ed39a-048e-4de0-812c-99daca774a99\") " pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.193944 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-logs" (OuterVolumeSpecName: "logs") pod "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" (UID: "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.196648 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" (UID: "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.199830 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-kube-api-access-gd2kg" (OuterVolumeSpecName: "kube-api-access-gd2kg") pod "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" (UID: "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"). InnerVolumeSpecName "kube-api-access-gd2kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.220018 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-scripts" (OuterVolumeSpecName: "scripts") pod "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" (UID: "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.263321 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028" (OuterVolumeSpecName: "glance") pod "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" (UID: "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"). InnerVolumeSpecName "pvc-de119cd6-b81e-445f-a9db-62090e36b028". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.279656 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea","Type":"ContainerDied","Data":"7e4587a2da0934a496a2efeb6f7e7c9883137ddfe15be35abf89aa9d8d24a099"} Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.285201 4706 scope.go:117] "RemoveContainer" containerID="0388e89c6d2f7c592342fa82247c63861f3a445ae230f096b01075abd5742a5a" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.288165 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.295066 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nkrt\" (UniqueName: \"kubernetes.io/projected/d05ed39a-048e-4de0-812c-99daca774a99-kube-api-access-8nkrt\") pod \"nova-cell1-2143-account-create-update-fqwzd\" (UID: \"d05ed39a-048e-4de0-812c-99daca774a99\") " pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.317842 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d05ed39a-048e-4de0-812c-99daca774a99-operator-scripts\") pod \"nova-cell1-2143-account-create-update-fqwzd\" (UID: \"d05ed39a-048e-4de0-812c-99daca774a99\") " pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.319019 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d05ed39a-048e-4de0-812c-99daca774a99-operator-scripts\") pod \"nova-cell1-2143-account-create-update-fqwzd\" (UID: \"d05ed39a-048e-4de0-812c-99daca774a99\") " pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.332673 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nkrt\" (UniqueName: \"kubernetes.io/projected/d05ed39a-048e-4de0-812c-99daca774a99-kube-api-access-8nkrt\") pod \"nova-cell1-2143-account-create-update-fqwzd\" (UID: \"d05ed39a-048e-4de0-812c-99daca774a99\") " pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.335961 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.336105 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") on node \"crc\" " Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.336188 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd2kg\" (UniqueName: \"kubernetes.io/projected/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-kube-api-access-gd2kg\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.336278 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.336367 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.369514 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" (UID: "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.421032 4706 scope.go:117] "RemoveContainer" containerID="acbdd1c35d016d77141180892bdb89076707fb1cb7d251b680a9452396dab032" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.438745 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" (UID: "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.440471 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.440778 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-de119cd6-b81e-445f-a9db-62090e36b028" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028") on node "crc" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.441879 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.441921 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.441934 4706 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.475672 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-config-data" (OuterVolumeSpecName: "config-data") pod "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" (UID: "8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.508543 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.544052 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:11 crc kubenswrapper[4706]: I1208 19:44:11.755110 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kd4rj"] Dec 08 19:44:11 crc kubenswrapper[4706]: W1208 19:44:11.770113 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ffa1493_bb68_48b6_8f13_c9c30358d1f3.slice/crio-55acd29c1be64b59ac4adb28024fe3085daff89f89db75c814438c979c3d18db WatchSource:0}: Error finding container 55acd29c1be64b59ac4adb28024fe3085daff89f89db75c814438c979c3d18db: Status 404 returned error can't find the container with id 55acd29c1be64b59ac4adb28024fe3085daff89f89db75c814438c979c3d18db Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.047023 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-c82mx"] Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.079503 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-t86b4"] Dec 08 19:44:12 crc kubenswrapper[4706]: W1208 19:44:12.157566 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod855d2f91_5dfa_401f_b1ac_a3f94287c245.slice/crio-6d4e081022b793572ce4c5db5d47cb062f9122f6051eefb2417a226e4a614414 WatchSource:0}: Error finding container 6d4e081022b793572ce4c5db5d47cb062f9122f6051eefb2417a226e4a614414: Status 404 returned error can't find the container with id 6d4e081022b793572ce4c5db5d47cb062f9122f6051eefb2417a226e4a614414 Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.295534 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c82mx" event={"ID":"50786407-6513-4d04-b13b-0367252b90af","Type":"ContainerStarted","Data":"f0d4b88c72efca82bf46af9de4f6e01f11b7b4009e22059ee0f2bc20c8aa4501"} Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.310681 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kd4rj" event={"ID":"3ffa1493-bb68-48b6-8f13-c9c30358d1f3","Type":"ContainerStarted","Data":"17832f12d8fd4aa04d54e49f062fac087c676ec1af1a1dbc40a930dc66c84c39"} Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.310900 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kd4rj" event={"ID":"3ffa1493-bb68-48b6-8f13-c9c30358d1f3","Type":"ContainerStarted","Data":"55acd29c1be64b59ac4adb28024fe3085daff89f89db75c814438c979c3d18db"} Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.312802 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-t86b4" event={"ID":"855d2f91-5dfa-401f-b1ac-a3f94287c245","Type":"ContainerStarted","Data":"6d4e081022b793572ce4c5db5d47cb062f9122f6051eefb2417a226e4a614414"} Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.349907 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-kd4rj" podStartSLOduration=2.349882203 podStartE2EDuration="2.349882203s" podCreationTimestamp="2025-12-08 19:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:44:12.32930591 +0000 UTC m=+1334.971506913" watchObservedRunningTime="2025-12-08 19:44:12.349882203 +0000 UTC m=+1334.992083206" Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.552637 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-758b-account-create-update-zh9x9"] Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.581002 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c56a-account-create-update-kl49q"] Dec 08 19:44:12 crc kubenswrapper[4706]: I1208 19:44:12.707346 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2143-account-create-update-fqwzd"] Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.326965 4706 generic.go:334] "Generic (PLEG): container finished" podID="3ffa1493-bb68-48b6-8f13-c9c30358d1f3" containerID="17832f12d8fd4aa04d54e49f062fac087c676ec1af1a1dbc40a930dc66c84c39" exitCode=0 Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.327071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kd4rj" event={"ID":"3ffa1493-bb68-48b6-8f13-c9c30358d1f3","Type":"ContainerDied","Data":"17832f12d8fd4aa04d54e49f062fac087c676ec1af1a1dbc40a930dc66c84c39"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.338688 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c56a-account-create-update-kl49q" event={"ID":"66571791-d210-477a-8951-20a0609193a9","Type":"ContainerStarted","Data":"1c626a309924b0d060b2ec6d4ee99df5f2fc5939f07d8b9caf0922aaf9202b18"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.338744 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c56a-account-create-update-kl49q" event={"ID":"66571791-d210-477a-8951-20a0609193a9","Type":"ContainerStarted","Data":"61893d4320bee9d68b19b234ca1bdcff5e62f9d89ef5922203c435c122bc322e"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.340744 4706 generic.go:334] "Generic (PLEG): container finished" podID="868f8ce6-c1f8-404d-8226-7fffb0518de4" containerID="092594639b173507d5ebac41d663c6f844dc9b6027306260eb017e99517ef3f8" exitCode=0 Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.340786 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-758b-account-create-update-zh9x9" event={"ID":"868f8ce6-c1f8-404d-8226-7fffb0518de4","Type":"ContainerDied","Data":"092594639b173507d5ebac41d663c6f844dc9b6027306260eb017e99517ef3f8"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.340802 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-758b-account-create-update-zh9x9" event={"ID":"868f8ce6-c1f8-404d-8226-7fffb0518de4","Type":"ContainerStarted","Data":"fcd8397a8d9dd691eb1f80e49ede6f99c786cc1980d9044865df3e1ff21b3b93"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.342882 4706 generic.go:334] "Generic (PLEG): container finished" podID="855d2f91-5dfa-401f-b1ac-a3f94287c245" containerID="ce88419453a5ffff15eead9387f7089800790b5d2b71518b90c308d1d45a1a36" exitCode=0 Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.342940 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-t86b4" event={"ID":"855d2f91-5dfa-401f-b1ac-a3f94287c245","Type":"ContainerDied","Data":"ce88419453a5ffff15eead9387f7089800790b5d2b71518b90c308d1d45a1a36"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.344626 4706 generic.go:334] "Generic (PLEG): container finished" podID="50786407-6513-4d04-b13b-0367252b90af" containerID="b0244d3b41d9204a8c97ca3de1cb3259469af58b87513fd4711ccd12fffdfe62" exitCode=0 Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.344713 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c82mx" event={"ID":"50786407-6513-4d04-b13b-0367252b90af","Type":"ContainerDied","Data":"b0244d3b41d9204a8c97ca3de1cb3259469af58b87513fd4711ccd12fffdfe62"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.346553 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" event={"ID":"d05ed39a-048e-4de0-812c-99daca774a99","Type":"ContainerStarted","Data":"b7900d848be946a04d35908bd11716c92feb0d454a870282e09895fe4d213f42"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.346580 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" event={"ID":"d05ed39a-048e-4de0-812c-99daca774a99","Type":"ContainerStarted","Data":"147f953a94b57f7755ae0a5533ce67c8bcaf87bf20881498d20751fa9a46b26e"} Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.378755 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-c56a-account-create-update-kl49q" podStartSLOduration=3.378721681 podStartE2EDuration="3.378721681s" podCreationTimestamp="2025-12-08 19:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:44:13.368184107 +0000 UTC m=+1336.010385110" watchObservedRunningTime="2025-12-08 19:44:13.378721681 +0000 UTC m=+1336.020922684" Dec 08 19:44:13 crc kubenswrapper[4706]: I1208 19:44:13.471898 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" podStartSLOduration=3.471875219 podStartE2EDuration="3.471875219s" podCreationTimestamp="2025-12-08 19:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:44:13.451543512 +0000 UTC m=+1336.093744535" watchObservedRunningTime="2025-12-08 19:44:13.471875219 +0000 UTC m=+1336.114076222" Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.310039 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.311573 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="ceilometer-central-agent" containerID="cri-o://57bf85b62b27f1ba5ee2aa0db4cee2def5dd29a65a5cc010e0583431398354cd" gracePeriod=30 Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.312288 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="ceilometer-notification-agent" containerID="cri-o://5808c7b997e14db62e6f8937c59569056dee8f68330e025b20e6890b2a9772c1" gracePeriod=30 Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.312332 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="sg-core" containerID="cri-o://608db59125990886f0f29367fcdd13d32c26ce2f6501e7eba2523008c9b7429c" gracePeriod=30 Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.312378 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="proxy-httpd" containerID="cri-o://0398181274faf2faa938930a494f942266d8782f3720622e205c4e7e3ee3a8f8" gracePeriod=30 Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.366141 4706 generic.go:334] "Generic (PLEG): container finished" podID="66571791-d210-477a-8951-20a0609193a9" containerID="1c626a309924b0d060b2ec6d4ee99df5f2fc5939f07d8b9caf0922aaf9202b18" exitCode=0 Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.366208 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c56a-account-create-update-kl49q" event={"ID":"66571791-d210-477a-8951-20a0609193a9","Type":"ContainerDied","Data":"1c626a309924b0d060b2ec6d4ee99df5f2fc5939f07d8b9caf0922aaf9202b18"} Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.370199 4706 generic.go:334] "Generic (PLEG): container finished" podID="d05ed39a-048e-4de0-812c-99daca774a99" containerID="b7900d848be946a04d35908bd11716c92feb0d454a870282e09895fe4d213f42" exitCode=0 Dec 08 19:44:14 crc kubenswrapper[4706]: I1208 19:44:14.370381 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" event={"ID":"d05ed39a-048e-4de0-812c-99daca774a99","Type":"ContainerDied","Data":"b7900d848be946a04d35908bd11716c92feb0d454a870282e09895fe4d213f42"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.103726 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.256294 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/855d2f91-5dfa-401f-b1ac-a3f94287c245-operator-scripts\") pod \"855d2f91-5dfa-401f-b1ac-a3f94287c245\" (UID: \"855d2f91-5dfa-401f-b1ac-a3f94287c245\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.257161 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv6hs\" (UniqueName: \"kubernetes.io/projected/855d2f91-5dfa-401f-b1ac-a3f94287c245-kube-api-access-sv6hs\") pod \"855d2f91-5dfa-401f-b1ac-a3f94287c245\" (UID: \"855d2f91-5dfa-401f-b1ac-a3f94287c245\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.264887 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/855d2f91-5dfa-401f-b1ac-a3f94287c245-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "855d2f91-5dfa-401f-b1ac-a3f94287c245" (UID: "855d2f91-5dfa-401f-b1ac-a3f94287c245"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.348734 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/855d2f91-5dfa-401f-b1ac-a3f94287c245-kube-api-access-sv6hs" (OuterVolumeSpecName: "kube-api-access-sv6hs") pod "855d2f91-5dfa-401f-b1ac-a3f94287c245" (UID: "855d2f91-5dfa-401f-b1ac-a3f94287c245"). InnerVolumeSpecName "kube-api-access-sv6hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.360139 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/855d2f91-5dfa-401f-b1ac-a3f94287c245-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.360191 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv6hs\" (UniqueName: \"kubernetes.io/projected/855d2f91-5dfa-401f-b1ac-a3f94287c245-kube-api-access-sv6hs\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.398066 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kd4rj" event={"ID":"3ffa1493-bb68-48b6-8f13-c9c30358d1f3","Type":"ContainerDied","Data":"55acd29c1be64b59ac4adb28024fe3085daff89f89db75c814438c979c3d18db"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.398116 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55acd29c1be64b59ac4adb28024fe3085daff89f89db75c814438c979c3d18db" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.409120 4706 generic.go:334] "Generic (PLEG): container finished" podID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerID="0398181274faf2faa938930a494f942266d8782f3720622e205c4e7e3ee3a8f8" exitCode=0 Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.409161 4706 generic.go:334] "Generic (PLEG): container finished" podID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerID="608db59125990886f0f29367fcdd13d32c26ce2f6501e7eba2523008c9b7429c" exitCode=2 Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.409173 4706 generic.go:334] "Generic (PLEG): container finished" podID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerID="5808c7b997e14db62e6f8937c59569056dee8f68330e025b20e6890b2a9772c1" exitCode=0 Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.409181 4706 generic.go:334] "Generic (PLEG): container finished" podID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerID="57bf85b62b27f1ba5ee2aa0db4cee2def5dd29a65a5cc010e0583431398354cd" exitCode=0 Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.409235 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerDied","Data":"0398181274faf2faa938930a494f942266d8782f3720622e205c4e7e3ee3a8f8"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.409283 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerDied","Data":"608db59125990886f0f29367fcdd13d32c26ce2f6501e7eba2523008c9b7429c"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.409296 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerDied","Data":"5808c7b997e14db62e6f8937c59569056dee8f68330e025b20e6890b2a9772c1"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.409307 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerDied","Data":"57bf85b62b27f1ba5ee2aa0db4cee2def5dd29a65a5cc010e0583431398354cd"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.418470 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-758b-account-create-update-zh9x9" event={"ID":"868f8ce6-c1f8-404d-8226-7fffb0518de4","Type":"ContainerDied","Data":"fcd8397a8d9dd691eb1f80e49ede6f99c786cc1980d9044865df3e1ff21b3b93"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.418526 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcd8397a8d9dd691eb1f80e49ede6f99c786cc1980d9044865df3e1ff21b3b93" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.425233 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-t86b4" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.425305 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-t86b4" event={"ID":"855d2f91-5dfa-401f-b1ac-a3f94287c245","Type":"ContainerDied","Data":"6d4e081022b793572ce4c5db5d47cb062f9122f6051eefb2417a226e4a614414"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.425357 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d4e081022b793572ce4c5db5d47cb062f9122f6051eefb2417a226e4a614414" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.428861 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c82mx" event={"ID":"50786407-6513-4d04-b13b-0367252b90af","Type":"ContainerDied","Data":"f0d4b88c72efca82bf46af9de4f6e01f11b7b4009e22059ee0f2bc20c8aa4501"} Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.428922 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0d4b88c72efca82bf46af9de4f6e01f11b7b4009e22059ee0f2bc20c8aa4501" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.525410 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.534829 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.542274 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.683154 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/868f8ce6-c1f8-404d-8226-7fffb0518de4-operator-scripts\") pod \"868f8ce6-c1f8-404d-8226-7fffb0518de4\" (UID: \"868f8ce6-c1f8-404d-8226-7fffb0518de4\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.683354 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-operator-scripts\") pod \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\" (UID: \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.683442 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f7v2\" (UniqueName: \"kubernetes.io/projected/868f8ce6-c1f8-404d-8226-7fffb0518de4-kube-api-access-7f7v2\") pod \"868f8ce6-c1f8-404d-8226-7fffb0518de4\" (UID: \"868f8ce6-c1f8-404d-8226-7fffb0518de4\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.683485 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29bwj\" (UniqueName: \"kubernetes.io/projected/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-kube-api-access-29bwj\") pod \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\" (UID: \"3ffa1493-bb68-48b6-8f13-c9c30358d1f3\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.683536 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50786407-6513-4d04-b13b-0367252b90af-operator-scripts\") pod \"50786407-6513-4d04-b13b-0367252b90af\" (UID: \"50786407-6513-4d04-b13b-0367252b90af\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.683645 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr85w\" (UniqueName: \"kubernetes.io/projected/50786407-6513-4d04-b13b-0367252b90af-kube-api-access-nr85w\") pod \"50786407-6513-4d04-b13b-0367252b90af\" (UID: \"50786407-6513-4d04-b13b-0367252b90af\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.686733 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50786407-6513-4d04-b13b-0367252b90af-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "50786407-6513-4d04-b13b-0367252b90af" (UID: "50786407-6513-4d04-b13b-0367252b90af"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.686851 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/868f8ce6-c1f8-404d-8226-7fffb0518de4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "868f8ce6-c1f8-404d-8226-7fffb0518de4" (UID: "868f8ce6-c1f8-404d-8226-7fffb0518de4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.688758 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3ffa1493-bb68-48b6-8f13-c9c30358d1f3" (UID: "3ffa1493-bb68-48b6-8f13-c9c30358d1f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.692287 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50786407-6513-4d04-b13b-0367252b90af-kube-api-access-nr85w" (OuterVolumeSpecName: "kube-api-access-nr85w") pod "50786407-6513-4d04-b13b-0367252b90af" (UID: "50786407-6513-4d04-b13b-0367252b90af"). InnerVolumeSpecName "kube-api-access-nr85w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.706714 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-kube-api-access-29bwj" (OuterVolumeSpecName: "kube-api-access-29bwj") pod "3ffa1493-bb68-48b6-8f13-c9c30358d1f3" (UID: "3ffa1493-bb68-48b6-8f13-c9c30358d1f3"). InnerVolumeSpecName "kube-api-access-29bwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.707991 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/868f8ce6-c1f8-404d-8226-7fffb0518de4-kube-api-access-7f7v2" (OuterVolumeSpecName: "kube-api-access-7f7v2") pod "868f8ce6-c1f8-404d-8226-7fffb0518de4" (UID: "868f8ce6-c1f8-404d-8226-7fffb0518de4"). InnerVolumeSpecName "kube-api-access-7f7v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.787376 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/868f8ce6-c1f8-404d-8226-7fffb0518de4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.787422 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.787438 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f7v2\" (UniqueName: \"kubernetes.io/projected/868f8ce6-c1f8-404d-8226-7fffb0518de4-kube-api-access-7f7v2\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.787455 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29bwj\" (UniqueName: \"kubernetes.io/projected/3ffa1493-bb68-48b6-8f13-c9c30358d1f3-kube-api-access-29bwj\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.787468 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50786407-6513-4d04-b13b-0367252b90af-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.787477 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr85w\" (UniqueName: \"kubernetes.io/projected/50786407-6513-4d04-b13b-0367252b90af-kube-api-access-nr85w\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.869755 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.991181 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-combined-ca-bundle\") pod \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.992118 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-config-data\") pod \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.992294 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-log-httpd\") pod \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.992327 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-sg-core-conf-yaml\") pod \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.992422 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-run-httpd\") pod \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.992515 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgr9v\" (UniqueName: \"kubernetes.io/projected/8f58e55a-69a4-457c-aecf-8681f88e0ce6-kube-api-access-tgr9v\") pod \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.992561 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-scripts\") pod \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\" (UID: \"8f58e55a-69a4-457c-aecf-8681f88e0ce6\") " Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.994467 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8f58e55a-69a4-457c-aecf-8681f88e0ce6" (UID: "8f58e55a-69a4-457c-aecf-8681f88e0ce6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:15 crc kubenswrapper[4706]: I1208 19:44:15.995110 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8f58e55a-69a4-457c-aecf-8681f88e0ce6" (UID: "8f58e55a-69a4-457c-aecf-8681f88e0ce6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.007981 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-scripts" (OuterVolumeSpecName: "scripts") pod "8f58e55a-69a4-457c-aecf-8681f88e0ce6" (UID: "8f58e55a-69a4-457c-aecf-8681f88e0ce6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.012456 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f58e55a-69a4-457c-aecf-8681f88e0ce6-kube-api-access-tgr9v" (OuterVolumeSpecName: "kube-api-access-tgr9v") pod "8f58e55a-69a4-457c-aecf-8681f88e0ce6" (UID: "8f58e55a-69a4-457c-aecf-8681f88e0ce6"). InnerVolumeSpecName "kube-api-access-tgr9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.060619 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8f58e55a-69a4-457c-aecf-8681f88e0ce6" (UID: "8f58e55a-69a4-457c-aecf-8681f88e0ce6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.096010 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.096048 4706 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.096062 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f58e55a-69a4-457c-aecf-8681f88e0ce6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.096073 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgr9v\" (UniqueName: \"kubernetes.io/projected/8f58e55a-69a4-457c-aecf-8681f88e0ce6-kube-api-access-tgr9v\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.096083 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.133707 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f58e55a-69a4-457c-aecf-8681f88e0ce6" (UID: "8f58e55a-69a4-457c-aecf-8681f88e0ce6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.157599 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-config-data" (OuterVolumeSpecName: "config-data") pod "8f58e55a-69a4-457c-aecf-8681f88e0ce6" (UID: "8f58e55a-69a4-457c-aecf-8681f88e0ce6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.175149 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.199609 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.199666 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f58e55a-69a4-457c-aecf-8681f88e0ce6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.245615 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.302130 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nkrt\" (UniqueName: \"kubernetes.io/projected/d05ed39a-048e-4de0-812c-99daca774a99-kube-api-access-8nkrt\") pod \"d05ed39a-048e-4de0-812c-99daca774a99\" (UID: \"d05ed39a-048e-4de0-812c-99daca774a99\") " Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.302315 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d05ed39a-048e-4de0-812c-99daca774a99-operator-scripts\") pod \"d05ed39a-048e-4de0-812c-99daca774a99\" (UID: \"d05ed39a-048e-4de0-812c-99daca774a99\") " Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.303713 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d05ed39a-048e-4de0-812c-99daca774a99-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d05ed39a-048e-4de0-812c-99daca774a99" (UID: "d05ed39a-048e-4de0-812c-99daca774a99"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.308742 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05ed39a-048e-4de0-812c-99daca774a99-kube-api-access-8nkrt" (OuterVolumeSpecName: "kube-api-access-8nkrt") pod "d05ed39a-048e-4de0-812c-99daca774a99" (UID: "d05ed39a-048e-4de0-812c-99daca774a99"). InnerVolumeSpecName "kube-api-access-8nkrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.336136 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.336493 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-log" containerID="cri-o://7c7a3136ddfe2fffec794311b2f8b777789da0e5fb2cba727560ec3e903035b5" gracePeriod=30 Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.336601 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-httpd" containerID="cri-o://51656795812e56dbc1e519431891b403757e847c1cc83f6cbc46b42c3550f121" gracePeriod=30 Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.404552 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66571791-d210-477a-8951-20a0609193a9-operator-scripts\") pod \"66571791-d210-477a-8951-20a0609193a9\" (UID: \"66571791-d210-477a-8951-20a0609193a9\") " Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.404674 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfv8b\" (UniqueName: \"kubernetes.io/projected/66571791-d210-477a-8951-20a0609193a9-kube-api-access-pfv8b\") pod \"66571791-d210-477a-8951-20a0609193a9\" (UID: \"66571791-d210-477a-8951-20a0609193a9\") " Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.405424 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66571791-d210-477a-8951-20a0609193a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "66571791-d210-477a-8951-20a0609193a9" (UID: "66571791-d210-477a-8951-20a0609193a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.405473 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nkrt\" (UniqueName: \"kubernetes.io/projected/d05ed39a-048e-4de0-812c-99daca774a99-kube-api-access-8nkrt\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.405499 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d05ed39a-048e-4de0-812c-99daca774a99-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.411810 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66571791-d210-477a-8951-20a0609193a9-kube-api-access-pfv8b" (OuterVolumeSpecName: "kube-api-access-pfv8b") pod "66571791-d210-477a-8951-20a0609193a9" (UID: "66571791-d210-477a-8951-20a0609193a9"). InnerVolumeSpecName "kube-api-access-pfv8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.452899 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f58e55a-69a4-457c-aecf-8681f88e0ce6","Type":"ContainerDied","Data":"357460b0c13d6d5602d31d7936eea15b973b4bb29858e883f6e57f9e48495936"} Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.452999 4706 scope.go:117] "RemoveContainer" containerID="0398181274faf2faa938930a494f942266d8782f3720622e205c4e7e3ee3a8f8" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.452937 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.458080 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c56a-account-create-update-kl49q" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.458401 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c56a-account-create-update-kl49q" event={"ID":"66571791-d210-477a-8951-20a0609193a9","Type":"ContainerDied","Data":"61893d4320bee9d68b19b234ca1bdcff5e62f9d89ef5922203c435c122bc322e"} Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.458463 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61893d4320bee9d68b19b234ca1bdcff5e62f9d89ef5922203c435c122bc322e" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.460753 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kd4rj" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.461167 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" event={"ID":"d05ed39a-048e-4de0-812c-99daca774a99","Type":"ContainerDied","Data":"147f953a94b57f7755ae0a5533ce67c8bcaf87bf20881498d20751fa9a46b26e"} Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.461218 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="147f953a94b57f7755ae0a5533ce67c8bcaf87bf20881498d20751fa9a46b26e" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.461306 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2143-account-create-update-fqwzd" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.461862 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c82mx" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.465455 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-758b-account-create-update-zh9x9" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.484625 4706 scope.go:117] "RemoveContainer" containerID="608db59125990886f0f29367fcdd13d32c26ce2f6501e7eba2523008c9b7429c" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.508402 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66571791-d210-477a-8951-20a0609193a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.508800 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfv8b\" (UniqueName: \"kubernetes.io/projected/66571791-d210-477a-8951-20a0609193a9-kube-api-access-pfv8b\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.547926 4706 scope.go:117] "RemoveContainer" containerID="5808c7b997e14db62e6f8937c59569056dee8f68330e025b20e6890b2a9772c1" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.587180 4706 scope.go:117] "RemoveContainer" containerID="57bf85b62b27f1ba5ee2aa0db4cee2def5dd29a65a5cc010e0583431398354cd" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.667383 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.700353 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.724029 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.724884 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="sg-core" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.724908 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="sg-core" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.724937 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerName="glance-httpd" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.724945 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerName="glance-httpd" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.724959 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="proxy-httpd" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.724966 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="proxy-httpd" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.724981 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855d2f91-5dfa-401f-b1ac-a3f94287c245" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.724988 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="855d2f91-5dfa-401f-b1ac-a3f94287c245" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.724999 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ffa1493-bb68-48b6-8f13-c9c30358d1f3" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725006 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ffa1493-bb68-48b6-8f13-c9c30358d1f3" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.725019 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="ceilometer-notification-agent" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725025 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="ceilometer-notification-agent" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.725039 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66571791-d210-477a-8951-20a0609193a9" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725046 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="66571791-d210-477a-8951-20a0609193a9" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.725056 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerName="glance-log" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725063 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerName="glance-log" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.725071 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="ceilometer-central-agent" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725078 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="ceilometer-central-agent" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.725097 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50786407-6513-4d04-b13b-0367252b90af" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725103 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="50786407-6513-4d04-b13b-0367252b90af" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.725116 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05ed39a-048e-4de0-812c-99daca774a99" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725123 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05ed39a-048e-4de0-812c-99daca774a99" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: E1208 19:44:16.725140 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="868f8ce6-c1f8-404d-8226-7fffb0518de4" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725147 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="868f8ce6-c1f8-404d-8226-7fffb0518de4" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725491 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerName="glance-log" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725502 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="ceilometer-notification-agent" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725535 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="sg-core" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725550 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="50786407-6513-4d04-b13b-0367252b90af" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725566 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="66571791-d210-477a-8951-20a0609193a9" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725574 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05ed39a-048e-4de0-812c-99daca774a99" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725585 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="proxy-httpd" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725615 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" containerName="glance-httpd" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725634 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="855d2f91-5dfa-401f-b1ac-a3f94287c245" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725643 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" containerName="ceilometer-central-agent" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725656 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="868f8ce6-c1f8-404d-8226-7fffb0518de4" containerName="mariadb-account-create-update" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.725665 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ffa1493-bb68-48b6-8f13-c9c30358d1f3" containerName="mariadb-database-create" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.729531 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.734931 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.739843 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.740395 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.830405 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.830474 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-config-data\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.830510 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhzgp\" (UniqueName: \"kubernetes.io/projected/d0182064-8d53-4dd6-9806-af7ae79b0162-kube-api-access-xhzgp\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.830546 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-run-httpd\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.830572 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-log-httpd\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.830600 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.830624 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-scripts\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.932596 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.932662 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-scripts\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.932820 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.932853 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-config-data\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.932878 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhzgp\" (UniqueName: \"kubernetes.io/projected/d0182064-8d53-4dd6-9806-af7ae79b0162-kube-api-access-xhzgp\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.932914 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-run-httpd\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.932943 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-log-httpd\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.933524 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-log-httpd\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.935001 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-run-httpd\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.944457 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.944694 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-config-data\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.945231 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.956120 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-scripts\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:16 crc kubenswrapper[4706]: I1208 19:44:16.971064 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhzgp\" (UniqueName: \"kubernetes.io/projected/d0182064-8d53-4dd6-9806-af7ae79b0162-kube-api-access-xhzgp\") pod \"ceilometer-0\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " pod="openstack/ceilometer-0" Dec 08 19:44:17 crc kubenswrapper[4706]: I1208 19:44:17.120929 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:17 crc kubenswrapper[4706]: I1208 19:44:17.504775 4706 generic.go:334] "Generic (PLEG): container finished" podID="e565173b-b249-4a74-b11e-92393ba53ac3" containerID="7c7a3136ddfe2fffec794311b2f8b777789da0e5fb2cba727560ec3e903035b5" exitCode=143 Dec 08 19:44:17 crc kubenswrapper[4706]: I1208 19:44:17.504908 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e565173b-b249-4a74-b11e-92393ba53ac3","Type":"ContainerDied","Data":"7c7a3136ddfe2fffec794311b2f8b777789da0e5fb2cba727560ec3e903035b5"} Dec 08 19:44:17 crc kubenswrapper[4706]: I1208 19:44:17.637504 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f58e55a-69a4-457c-aecf-8681f88e0ce6" path="/var/lib/kubelet/pods/8f58e55a-69a4-457c-aecf-8681f88e0ce6/volumes" Dec 08 19:44:17 crc kubenswrapper[4706]: W1208 19:44:17.784641 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0182064_8d53_4dd6_9806_af7ae79b0162.slice/crio-cc0e14ad20561d128acac0e4ac043738eeda169b8e7affa6b7dc8310f95dd8db WatchSource:0}: Error finding container cc0e14ad20561d128acac0e4ac043738eeda169b8e7affa6b7dc8310f95dd8db: Status 404 returned error can't find the container with id cc0e14ad20561d128acac0e4ac043738eeda169b8e7affa6b7dc8310f95dd8db Dec 08 19:44:17 crc kubenswrapper[4706]: I1208 19:44:17.790696 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:18 crc kubenswrapper[4706]: I1208 19:44:18.118784 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:18 crc kubenswrapper[4706]: I1208 19:44:18.540903 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerStarted","Data":"cc0e14ad20561d128acac0e4ac043738eeda169b8e7affa6b7dc8310f95dd8db"} Dec 08 19:44:19 crc kubenswrapper[4706]: I1208 19:44:19.555330 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerStarted","Data":"efed7ada307f4754f7917b3220c586d6e420f0ba3f503c0b7bf0034a8981a879"} Dec 08 19:44:19 crc kubenswrapper[4706]: I1208 19:44:19.556027 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerStarted","Data":"2b6247a522967b0b85902217ebdf2c8ff48bebed9ea74103fcfebc9cfb37ee3b"} Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.006018 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.184:9292/healthcheck\": read tcp 10.217.0.2:41196->10.217.0.184:9292: read: connection reset by peer" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.006672 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.184:9292/healthcheck\": read tcp 10.217.0.2:41210->10.217.0.184:9292: read: connection reset by peer" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.582887 4706 generic.go:334] "Generic (PLEG): container finished" podID="e565173b-b249-4a74-b11e-92393ba53ac3" containerID="51656795812e56dbc1e519431891b403757e847c1cc83f6cbc46b42c3550f121" exitCode=0 Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.582939 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e565173b-b249-4a74-b11e-92393ba53ac3","Type":"ContainerDied","Data":"51656795812e56dbc1e519431891b403757e847c1cc83f6cbc46b42c3550f121"} Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.593136 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerStarted","Data":"969fb54e687f7c9bbe8a5e16d6e6227eb18afa8ef6facdd9f962d0f280e92412"} Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.739545 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.836128 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-combined-ca-bundle\") pod \"e565173b-b249-4a74-b11e-92393ba53ac3\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.836381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"e565173b-b249-4a74-b11e-92393ba53ac3\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.836502 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-logs\") pod \"e565173b-b249-4a74-b11e-92393ba53ac3\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.836547 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-scripts\") pod \"e565173b-b249-4a74-b11e-92393ba53ac3\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.836626 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-httpd-run\") pod \"e565173b-b249-4a74-b11e-92393ba53ac3\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.836697 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkg78\" (UniqueName: \"kubernetes.io/projected/e565173b-b249-4a74-b11e-92393ba53ac3-kube-api-access-bkg78\") pod \"e565173b-b249-4a74-b11e-92393ba53ac3\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.836737 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-config-data\") pod \"e565173b-b249-4a74-b11e-92393ba53ac3\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.836835 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-internal-tls-certs\") pod \"e565173b-b249-4a74-b11e-92393ba53ac3\" (UID: \"e565173b-b249-4a74-b11e-92393ba53ac3\") " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.837106 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e565173b-b249-4a74-b11e-92393ba53ac3" (UID: "e565173b-b249-4a74-b11e-92393ba53ac3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.837200 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-logs" (OuterVolumeSpecName: "logs") pod "e565173b-b249-4a74-b11e-92393ba53ac3" (UID: "e565173b-b249-4a74-b11e-92393ba53ac3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.838127 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.838141 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e565173b-b249-4a74-b11e-92393ba53ac3-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.853489 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e565173b-b249-4a74-b11e-92393ba53ac3-kube-api-access-bkg78" (OuterVolumeSpecName: "kube-api-access-bkg78") pod "e565173b-b249-4a74-b11e-92393ba53ac3" (UID: "e565173b-b249-4a74-b11e-92393ba53ac3"). InnerVolumeSpecName "kube-api-access-bkg78". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.877992 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-scripts" (OuterVolumeSpecName: "scripts") pod "e565173b-b249-4a74-b11e-92393ba53ac3" (UID: "e565173b-b249-4a74-b11e-92393ba53ac3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.900035 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7" (OuterVolumeSpecName: "glance") pod "e565173b-b249-4a74-b11e-92393ba53ac3" (UID: "e565173b-b249-4a74-b11e-92393ba53ac3"). InnerVolumeSpecName "pvc-157f530e-9886-42dc-a031-b8a66480e0f7". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.902786 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e565173b-b249-4a74-b11e-92393ba53ac3" (UID: "e565173b-b249-4a74-b11e-92393ba53ac3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.943867 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkg78\" (UniqueName: \"kubernetes.io/projected/e565173b-b249-4a74-b11e-92393ba53ac3-kube-api-access-bkg78\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.943909 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.943950 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") on node \"crc\" " Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.943962 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.964580 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e565173b-b249-4a74-b11e-92393ba53ac3" (UID: "e565173b-b249-4a74-b11e-92393ba53ac3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.969499 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-config-data" (OuterVolumeSpecName: "config-data") pod "e565173b-b249-4a74-b11e-92393ba53ac3" (UID: "e565173b-b249-4a74-b11e-92393ba53ac3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.993456 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 08 19:44:20 crc kubenswrapper[4706]: I1208 19:44:20.993657 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-157f530e-9886-42dc-a031-b8a66480e0f7" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7") on node "crc" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.046062 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.046132 4706 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e565173b-b249-4a74-b11e-92393ba53ac3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.046147 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.180494 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6qpgg"] Dec 08 19:44:21 crc kubenswrapper[4706]: E1208 19:44:21.181018 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-httpd" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.181059 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-httpd" Dec 08 19:44:21 crc kubenswrapper[4706]: E1208 19:44:21.181090 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-log" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.181098 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-log" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.181322 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-httpd" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.181350 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" containerName="glance-log" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.182567 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.186957 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.187251 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.187449 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-c2cjl" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.216303 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6qpgg"] Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.256172 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.256552 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-scripts\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.256685 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-config-data\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.256738 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd98j\" (UniqueName: \"kubernetes.io/projected/e8823e38-e8cd-47b1-8408-b999c691f189-kube-api-access-xd98j\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.359166 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.359796 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-scripts\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.359861 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-config-data\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.359896 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd98j\" (UniqueName: \"kubernetes.io/projected/e8823e38-e8cd-47b1-8408-b999c691f189-kube-api-access-xd98j\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.365678 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-scripts\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.365859 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.365924 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-config-data\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.382505 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd98j\" (UniqueName: \"kubernetes.io/projected/e8823e38-e8cd-47b1-8408-b999c691f189-kube-api-access-xd98j\") pod \"nova-cell0-conductor-db-sync-6qpgg\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.546558 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.673764 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e565173b-b249-4a74-b11e-92393ba53ac3","Type":"ContainerDied","Data":"721e7d401fd2d51024d9552f82bb5a23620971cd18aee143c874a5c0c361d3b0"} Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.673893 4706 scope.go:117] "RemoveContainer" containerID="51656795812e56dbc1e519431891b403757e847c1cc83f6cbc46b42c3550f121" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.673924 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.733352 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.757231 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.794856 4706 scope.go:117] "RemoveContainer" containerID="7c7a3136ddfe2fffec794311b2f8b777789da0e5fb2cba727560ec3e903035b5" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.808351 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.810788 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.821958 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.822684 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.822684 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.823052 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kvdqs" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.823130 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.876649 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6249f0fa-1385-4546-89c3-60a91e24da43-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.876816 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.876910 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6249f0fa-1385-4546-89c3-60a91e24da43-logs\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.877057 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.877097 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92jhq\" (UniqueName: \"kubernetes.io/projected/6249f0fa-1385-4546-89c3-60a91e24da43-kube-api-access-92jhq\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.877213 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.877295 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.877385 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.979857 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.979941 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.980003 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.980123 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6249f0fa-1385-4546-89c3-60a91e24da43-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.980153 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.980198 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6249f0fa-1385-4546-89c3-60a91e24da43-logs\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.980234 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.980256 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92jhq\" (UniqueName: \"kubernetes.io/projected/6249f0fa-1385-4546-89c3-60a91e24da43-kube-api-access-92jhq\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.984743 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6249f0fa-1385-4546-89c3-60a91e24da43-logs\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.988229 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6249f0fa-1385-4546-89c3-60a91e24da43-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.995462 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.998337 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.998856 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.999088 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:44:21 crc kubenswrapper[4706]: I1208 19:44:21.999114 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b040d665527e04c44fa620444bb13c4fde426c5d08f49faa6c04df54fad7a449/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.002015 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92jhq\" (UniqueName: \"kubernetes.io/projected/6249f0fa-1385-4546-89c3-60a91e24da43-kube-api-access-92jhq\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.002634 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6249f0fa-1385-4546-89c3-60a91e24da43-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.069401 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-157f530e-9886-42dc-a031-b8a66480e0f7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-157f530e-9886-42dc-a031-b8a66480e0f7\") pod \"glance-default-internal-api-0\" (UID: \"6249f0fa-1385-4546-89c3-60a91e24da43\") " pod="openstack/glance-default-internal-api-0" Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.156297 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.267199 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6qpgg"] Dec 08 19:44:22 crc kubenswrapper[4706]: W1208 19:44:22.279428 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8823e38_e8cd_47b1_8408_b999c691f189.slice/crio-bcdcdb1aeaadd0e534e57d8ac2322630de7bc8c0b5152a6a83aa82fedd064495 WatchSource:0}: Error finding container bcdcdb1aeaadd0e534e57d8ac2322630de7bc8c0b5152a6a83aa82fedd064495: Status 404 returned error can't find the container with id bcdcdb1aeaadd0e534e57d8ac2322630de7bc8c0b5152a6a83aa82fedd064495 Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.772991 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" event={"ID":"e8823e38-e8cd-47b1-8408-b999c691f189","Type":"ContainerStarted","Data":"bcdcdb1aeaadd0e534e57d8ac2322630de7bc8c0b5152a6a83aa82fedd064495"} Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.798716 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerStarted","Data":"c5c825fd2e60cef904a0506aa1be8b12e39cc7ebcef38b7ffdca7f42efcace46"} Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.799009 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="ceilometer-central-agent" containerID="cri-o://2b6247a522967b0b85902217ebdf2c8ff48bebed9ea74103fcfebc9cfb37ee3b" gracePeriod=30 Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.799448 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.799585 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="proxy-httpd" containerID="cri-o://c5c825fd2e60cef904a0506aa1be8b12e39cc7ebcef38b7ffdca7f42efcace46" gracePeriod=30 Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.799828 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="sg-core" containerID="cri-o://969fb54e687f7c9bbe8a5e16d6e6227eb18afa8ef6facdd9f962d0f280e92412" gracePeriod=30 Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.799910 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="ceilometer-notification-agent" containerID="cri-o://efed7ada307f4754f7917b3220c586d6e420f0ba3f503c0b7bf0034a8981a879" gracePeriod=30 Dec 08 19:44:22 crc kubenswrapper[4706]: I1208 19:44:22.843304 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.061073465 podStartE2EDuration="6.843255957s" podCreationTimestamp="2025-12-08 19:44:16 +0000 UTC" firstStartedPulling="2025-12-08 19:44:17.786994564 +0000 UTC m=+1340.429195557" lastFinishedPulling="2025-12-08 19:44:21.569177036 +0000 UTC m=+1344.211378049" observedRunningTime="2025-12-08 19:44:22.840976282 +0000 UTC m=+1345.483177295" watchObservedRunningTime="2025-12-08 19:44:22.843255957 +0000 UTC m=+1345.485456960" Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.044301 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.641301 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e565173b-b249-4a74-b11e-92393ba53ac3" path="/var/lib/kubelet/pods/e565173b-b249-4a74-b11e-92393ba53ac3/volumes" Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.826573 4706 generic.go:334] "Generic (PLEG): container finished" podID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerID="c5c825fd2e60cef904a0506aa1be8b12e39cc7ebcef38b7ffdca7f42efcace46" exitCode=0 Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.826619 4706 generic.go:334] "Generic (PLEG): container finished" podID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerID="969fb54e687f7c9bbe8a5e16d6e6227eb18afa8ef6facdd9f962d0f280e92412" exitCode=2 Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.826628 4706 generic.go:334] "Generic (PLEG): container finished" podID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerID="efed7ada307f4754f7917b3220c586d6e420f0ba3f503c0b7bf0034a8981a879" exitCode=0 Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.826661 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerDied","Data":"c5c825fd2e60cef904a0506aa1be8b12e39cc7ebcef38b7ffdca7f42efcace46"} Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.826701 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerDied","Data":"969fb54e687f7c9bbe8a5e16d6e6227eb18afa8ef6facdd9f962d0f280e92412"} Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.826716 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerDied","Data":"efed7ada307f4754f7917b3220c586d6e420f0ba3f503c0b7bf0034a8981a879"} Dec 08 19:44:23 crc kubenswrapper[4706]: I1208 19:44:23.828336 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6249f0fa-1385-4546-89c3-60a91e24da43","Type":"ContainerStarted","Data":"b88b692c1db126173ecf536cf8c6b72617c933c7a37d034e67c551c9c8afebef"} Dec 08 19:44:24 crc kubenswrapper[4706]: I1208 19:44:24.856172 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6249f0fa-1385-4546-89c3-60a91e24da43","Type":"ContainerStarted","Data":"d256ff268c9cc53ad038c9f8b162a3d810c4ffa271a3122e762d7fdb3f4eb5b8"} Dec 08 19:44:24 crc kubenswrapper[4706]: I1208 19:44:24.857038 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6249f0fa-1385-4546-89c3-60a91e24da43","Type":"ContainerStarted","Data":"3c90123cba79e329296ca2a02372c29992a09159cd510e054d6f18fb3839f694"} Dec 08 19:44:24 crc kubenswrapper[4706]: I1208 19:44:24.885704 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.885661104 podStartE2EDuration="3.885661104s" podCreationTimestamp="2025-12-08 19:44:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:44:24.879408523 +0000 UTC m=+1347.521609526" watchObservedRunningTime="2025-12-08 19:44:24.885661104 +0000 UTC m=+1347.527862107" Dec 08 19:44:27 crc kubenswrapper[4706]: I1208 19:44:27.909931 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerDied","Data":"2b6247a522967b0b85902217ebdf2c8ff48bebed9ea74103fcfebc9cfb37ee3b"} Dec 08 19:44:27 crc kubenswrapper[4706]: I1208 19:44:27.909932 4706 generic.go:334] "Generic (PLEG): container finished" podID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerID="2b6247a522967b0b85902217ebdf2c8ff48bebed9ea74103fcfebc9cfb37ee3b" exitCode=0 Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.157336 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.158014 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.232906 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.233345 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.691723 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.805459 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-config-data\") pod \"d0182064-8d53-4dd6-9806-af7ae79b0162\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.805599 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhzgp\" (UniqueName: \"kubernetes.io/projected/d0182064-8d53-4dd6-9806-af7ae79b0162-kube-api-access-xhzgp\") pod \"d0182064-8d53-4dd6-9806-af7ae79b0162\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.805695 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-log-httpd\") pod \"d0182064-8d53-4dd6-9806-af7ae79b0162\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.805770 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-run-httpd\") pod \"d0182064-8d53-4dd6-9806-af7ae79b0162\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.805843 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-sg-core-conf-yaml\") pod \"d0182064-8d53-4dd6-9806-af7ae79b0162\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.805939 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-scripts\") pod \"d0182064-8d53-4dd6-9806-af7ae79b0162\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.806031 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-combined-ca-bundle\") pod \"d0182064-8d53-4dd6-9806-af7ae79b0162\" (UID: \"d0182064-8d53-4dd6-9806-af7ae79b0162\") " Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.806791 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d0182064-8d53-4dd6-9806-af7ae79b0162" (UID: "d0182064-8d53-4dd6-9806-af7ae79b0162"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.812078 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d0182064-8d53-4dd6-9806-af7ae79b0162" (UID: "d0182064-8d53-4dd6-9806-af7ae79b0162"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.814508 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0182064-8d53-4dd6-9806-af7ae79b0162-kube-api-access-xhzgp" (OuterVolumeSpecName: "kube-api-access-xhzgp") pod "d0182064-8d53-4dd6-9806-af7ae79b0162" (UID: "d0182064-8d53-4dd6-9806-af7ae79b0162"). InnerVolumeSpecName "kube-api-access-xhzgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.815691 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-scripts" (OuterVolumeSpecName: "scripts") pod "d0182064-8d53-4dd6-9806-af7ae79b0162" (UID: "d0182064-8d53-4dd6-9806-af7ae79b0162"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.846493 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d0182064-8d53-4dd6-9806-af7ae79b0162" (UID: "d0182064-8d53-4dd6-9806-af7ae79b0162"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.909092 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0182064-8d53-4dd6-9806-af7ae79b0162" (UID: "d0182064-8d53-4dd6-9806-af7ae79b0162"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.909869 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhzgp\" (UniqueName: \"kubernetes.io/projected/d0182064-8d53-4dd6-9806-af7ae79b0162-kube-api-access-xhzgp\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.909908 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.909922 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0182064-8d53-4dd6-9806-af7ae79b0162-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.909937 4706 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.909950 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.909964 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.920422 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-config-data" (OuterVolumeSpecName: "config-data") pod "d0182064-8d53-4dd6-9806-af7ae79b0162" (UID: "d0182064-8d53-4dd6-9806-af7ae79b0162"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.975027 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" event={"ID":"e8823e38-e8cd-47b1-8408-b999c691f189","Type":"ContainerStarted","Data":"bd2970d9cf0ca99e8844cccb6d28536e32c753aae0789f12d236a6fec646977f"} Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.978417 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0182064-8d53-4dd6-9806-af7ae79b0162","Type":"ContainerDied","Data":"cc0e14ad20561d128acac0e4ac043738eeda169b8e7affa6b7dc8310f95dd8db"} Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.978478 4706 scope.go:117] "RemoveContainer" containerID="c5c825fd2e60cef904a0506aa1be8b12e39cc7ebcef38b7ffdca7f42efcace46" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.978533 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.978746 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:32 crc kubenswrapper[4706]: I1208 19:44:32.979250 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.006238 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" podStartSLOduration=2.005536468 podStartE2EDuration="12.00620764s" podCreationTimestamp="2025-12-08 19:44:21 +0000 UTC" firstStartedPulling="2025-12-08 19:44:22.284489099 +0000 UTC m=+1344.926690102" lastFinishedPulling="2025-12-08 19:44:32.285160271 +0000 UTC m=+1354.927361274" observedRunningTime="2025-12-08 19:44:32.995375288 +0000 UTC m=+1355.637576291" watchObservedRunningTime="2025-12-08 19:44:33.00620764 +0000 UTC m=+1355.648408643" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.012051 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0182064-8d53-4dd6-9806-af7ae79b0162-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.012554 4706 scope.go:117] "RemoveContainer" containerID="969fb54e687f7c9bbe8a5e16d6e6227eb18afa8ef6facdd9f962d0f280e92412" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.036614 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.050783 4706 scope.go:117] "RemoveContainer" containerID="efed7ada307f4754f7917b3220c586d6e420f0ba3f503c0b7bf0034a8981a879" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.057336 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.075892 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:33 crc kubenswrapper[4706]: E1208 19:44:33.076625 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="ceilometer-notification-agent" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.076657 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="ceilometer-notification-agent" Dec 08 19:44:33 crc kubenswrapper[4706]: E1208 19:44:33.076674 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="ceilometer-central-agent" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.076684 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="ceilometer-central-agent" Dec 08 19:44:33 crc kubenswrapper[4706]: E1208 19:44:33.076701 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="proxy-httpd" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.076710 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="proxy-httpd" Dec 08 19:44:33 crc kubenswrapper[4706]: E1208 19:44:33.076738 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="sg-core" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.076747 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="sg-core" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.077064 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="ceilometer-notification-agent" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.077092 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="sg-core" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.077116 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="proxy-httpd" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.077131 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" containerName="ceilometer-central-agent" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.079794 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.084649 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.089500 4706 scope.go:117] "RemoveContainer" containerID="2b6247a522967b0b85902217ebdf2c8ff48bebed9ea74103fcfebc9cfb37ee3b" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.100254 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.100578 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.223973 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.224054 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-log-httpd\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.224128 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-run-httpd\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.224182 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.224221 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqnlb\" (UniqueName: \"kubernetes.io/projected/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-kube-api-access-tqnlb\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.224292 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-config-data\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.224347 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-scripts\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.328371 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.328788 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqnlb\" (UniqueName: \"kubernetes.io/projected/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-kube-api-access-tqnlb\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.328896 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-config-data\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.329008 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-scripts\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.329147 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.329247 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-log-httpd\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.329415 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-run-httpd\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.330373 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-run-httpd\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.330690 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-log-httpd\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.345460 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.348710 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-config-data\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.351783 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.380144 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqnlb\" (UniqueName: \"kubernetes.io/projected/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-kube-api-access-tqnlb\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.398403 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-scripts\") pod \"ceilometer-0\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.423902 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:44:33 crc kubenswrapper[4706]: I1208 19:44:33.632733 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0182064-8d53-4dd6-9806-af7ae79b0162" path="/var/lib/kubelet/pods/d0182064-8d53-4dd6-9806-af7ae79b0162/volumes" Dec 08 19:44:34 crc kubenswrapper[4706]: I1208 19:44:34.058780 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:44:35 crc kubenswrapper[4706]: I1208 19:44:35.024142 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerStarted","Data":"17b0b6bcc14fc233c319dad0dfdcdff0012767ebde787433b0dc828c158eb113"} Dec 08 19:44:35 crc kubenswrapper[4706]: I1208 19:44:35.848658 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:35 crc kubenswrapper[4706]: I1208 19:44:35.849326 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:44:35 crc kubenswrapper[4706]: I1208 19:44:35.858788 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 08 19:44:36 crc kubenswrapper[4706]: I1208 19:44:36.059210 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerStarted","Data":"9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef"} Dec 08 19:44:37 crc kubenswrapper[4706]: I1208 19:44:37.096308 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerStarted","Data":"23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940"} Dec 08 19:44:37 crc kubenswrapper[4706]: I1208 19:44:37.096791 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerStarted","Data":"f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537"} Dec 08 19:44:38 crc kubenswrapper[4706]: I1208 19:44:38.116108 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerStarted","Data":"68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5"} Dec 08 19:44:38 crc kubenswrapper[4706]: I1208 19:44:38.116760 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 19:44:38 crc kubenswrapper[4706]: I1208 19:44:38.147578 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.491574708 podStartE2EDuration="5.147548509s" podCreationTimestamp="2025-12-08 19:44:33 +0000 UTC" firstStartedPulling="2025-12-08 19:44:34.063574262 +0000 UTC m=+1356.705775265" lastFinishedPulling="2025-12-08 19:44:37.719548063 +0000 UTC m=+1360.361749066" observedRunningTime="2025-12-08 19:44:38.146148969 +0000 UTC m=+1360.788349982" watchObservedRunningTime="2025-12-08 19:44:38.147548509 +0000 UTC m=+1360.789749512" Dec 08 19:44:41 crc kubenswrapper[4706]: I1208 19:44:41.821872 4706 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea] : Timed out while waiting for systemd to remove kubepods-besteffort-pod8ff7c9db_1bb9_48b8_a7ac_b0468bda7fea.slice" Dec 08 19:44:41 crc kubenswrapper[4706]: E1208 19:44:41.822537 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea] : Timed out while waiting for systemd to remove kubepods-besteffort-pod8ff7c9db_1bb9_48b8_a7ac_b0468bda7fea.slice" pod="openstack/glance-default-external-api-0" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.171670 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.210568 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.236710 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.250337 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.252953 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.262802 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.263485 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.302040 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.377093 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-logs\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.377525 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.377573 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-config-data\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.377637 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-scripts\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.377708 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2mw8\" (UniqueName: \"kubernetes.io/projected/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-kube-api-access-t2mw8\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.377793 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.377833 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.377892 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.479813 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.479876 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.479923 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.479983 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-logs\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.480009 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.480038 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-config-data\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.480084 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-scripts\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.480132 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2mw8\" (UniqueName: \"kubernetes.io/projected/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-kube-api-access-t2mw8\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.481087 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.486660 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-logs\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.509774 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.530358 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.530681 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/67f02492c48f5b6b986fa4270f8226cd0203948331bc87047321fea71b95adc3/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.556368 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-config-data\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.557288 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-scripts\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.558000 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.558148 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2mw8\" (UniqueName: \"kubernetes.io/projected/8424f86e-b640-44f6-abd9-dc8c7a5a55cc-kube-api-access-t2mw8\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.681549 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-de119cd6-b81e-445f-a9db-62090e36b028\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de119cd6-b81e-445f-a9db-62090e36b028\") pod \"glance-default-external-api-0\" (UID: \"8424f86e-b640-44f6-abd9-dc8c7a5a55cc\") " pod="openstack/glance-default-external-api-0" Dec 08 19:44:42 crc kubenswrapper[4706]: I1208 19:44:42.877736 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 19:44:43 crc kubenswrapper[4706]: I1208 19:44:43.622062 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea" path="/var/lib/kubelet/pods/8ff7c9db-1bb9-48b8-a7ac-b0468bda7fea/volumes" Dec 08 19:44:43 crc kubenswrapper[4706]: I1208 19:44:43.625807 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 19:44:43 crc kubenswrapper[4706]: W1208 19:44:43.638209 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8424f86e_b640_44f6_abd9_dc8c7a5a55cc.slice/crio-d6e203fc0e476c826a23c79ceae152ba6b784d6600eeab8011ed3aec8c532d92 WatchSource:0}: Error finding container d6e203fc0e476c826a23c79ceae152ba6b784d6600eeab8011ed3aec8c532d92: Status 404 returned error can't find the container with id d6e203fc0e476c826a23c79ceae152ba6b784d6600eeab8011ed3aec8c532d92 Dec 08 19:44:44 crc kubenswrapper[4706]: I1208 19:44:44.202103 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8424f86e-b640-44f6-abd9-dc8c7a5a55cc","Type":"ContainerStarted","Data":"d6e203fc0e476c826a23c79ceae152ba6b784d6600eeab8011ed3aec8c532d92"} Dec 08 19:44:45 crc kubenswrapper[4706]: I1208 19:44:45.214378 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8424f86e-b640-44f6-abd9-dc8c7a5a55cc","Type":"ContainerStarted","Data":"3c5ff013979f42242999601607657b058e742fecb1a46a1fc0cb2f42e8eb2c63"} Dec 08 19:44:45 crc kubenswrapper[4706]: I1208 19:44:45.214743 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8424f86e-b640-44f6-abd9-dc8c7a5a55cc","Type":"ContainerStarted","Data":"cb148e01201ba636319d6db6ed36c54ad6c60f169ecc228f3dc71d63b7a40453"} Dec 08 19:44:45 crc kubenswrapper[4706]: I1208 19:44:45.254023 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.253992903 podStartE2EDuration="3.253992903s" podCreationTimestamp="2025-12-08 19:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:44:45.247037172 +0000 UTC m=+1367.889238175" watchObservedRunningTime="2025-12-08 19:44:45.253992903 +0000 UTC m=+1367.896193906" Dec 08 19:44:47 crc kubenswrapper[4706]: I1208 19:44:47.239173 4706 generic.go:334] "Generic (PLEG): container finished" podID="e8823e38-e8cd-47b1-8408-b999c691f189" containerID="bd2970d9cf0ca99e8844cccb6d28536e32c753aae0789f12d236a6fec646977f" exitCode=0 Dec 08 19:44:47 crc kubenswrapper[4706]: I1208 19:44:47.239228 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" event={"ID":"e8823e38-e8cd-47b1-8408-b999c691f189","Type":"ContainerDied","Data":"bd2970d9cf0ca99e8844cccb6d28536e32c753aae0789f12d236a6fec646977f"} Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.789655 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.883090 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-scripts\") pod \"e8823e38-e8cd-47b1-8408-b999c691f189\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.884605 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd98j\" (UniqueName: \"kubernetes.io/projected/e8823e38-e8cd-47b1-8408-b999c691f189-kube-api-access-xd98j\") pod \"e8823e38-e8cd-47b1-8408-b999c691f189\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.884769 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-combined-ca-bundle\") pod \"e8823e38-e8cd-47b1-8408-b999c691f189\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.907256 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-scripts" (OuterVolumeSpecName: "scripts") pod "e8823e38-e8cd-47b1-8408-b999c691f189" (UID: "e8823e38-e8cd-47b1-8408-b999c691f189"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.909299 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8823e38-e8cd-47b1-8408-b999c691f189-kube-api-access-xd98j" (OuterVolumeSpecName: "kube-api-access-xd98j") pod "e8823e38-e8cd-47b1-8408-b999c691f189" (UID: "e8823e38-e8cd-47b1-8408-b999c691f189"). InnerVolumeSpecName "kube-api-access-xd98j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.956377 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8823e38-e8cd-47b1-8408-b999c691f189" (UID: "e8823e38-e8cd-47b1-8408-b999c691f189"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.986920 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-config-data\") pod \"e8823e38-e8cd-47b1-8408-b999c691f189\" (UID: \"e8823e38-e8cd-47b1-8408-b999c691f189\") " Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.987569 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.987594 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:48 crc kubenswrapper[4706]: I1208 19:44:48.987609 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd98j\" (UniqueName: \"kubernetes.io/projected/e8823e38-e8cd-47b1-8408-b999c691f189-kube-api-access-xd98j\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.040281 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-config-data" (OuterVolumeSpecName: "config-data") pod "e8823e38-e8cd-47b1-8408-b999c691f189" (UID: "e8823e38-e8cd-47b1-8408-b999c691f189"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.088972 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8823e38-e8cd-47b1-8408-b999c691f189-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.267851 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" event={"ID":"e8823e38-e8cd-47b1-8408-b999c691f189","Type":"ContainerDied","Data":"bcdcdb1aeaadd0e534e57d8ac2322630de7bc8c0b5152a6a83aa82fedd064495"} Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.267905 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcdcdb1aeaadd0e534e57d8ac2322630de7bc8c0b5152a6a83aa82fedd064495" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.267958 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6qpgg" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.398960 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 08 19:44:49 crc kubenswrapper[4706]: E1208 19:44:49.399870 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8823e38-e8cd-47b1-8408-b999c691f189" containerName="nova-cell0-conductor-db-sync" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.399896 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8823e38-e8cd-47b1-8408-b999c691f189" containerName="nova-cell0-conductor-db-sync" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.400173 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8823e38-e8cd-47b1-8408-b999c691f189" containerName="nova-cell0-conductor-db-sync" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.401292 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.404009 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-c2cjl" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.404069 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.430689 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.500493 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667cc0c8-9292-4cbd-a013-bd645c4b57d7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.500597 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667cc0c8-9292-4cbd-a013-bd645c4b57d7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.500774 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvpzr\" (UniqueName: \"kubernetes.io/projected/667cc0c8-9292-4cbd-a013-bd645c4b57d7-kube-api-access-xvpzr\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.603045 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvpzr\" (UniqueName: \"kubernetes.io/projected/667cc0c8-9292-4cbd-a013-bd645c4b57d7-kube-api-access-xvpzr\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.603188 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667cc0c8-9292-4cbd-a013-bd645c4b57d7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.603252 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667cc0c8-9292-4cbd-a013-bd645c4b57d7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.608325 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667cc0c8-9292-4cbd-a013-bd645c4b57d7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.608891 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667cc0c8-9292-4cbd-a013-bd645c4b57d7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.620060 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvpzr\" (UniqueName: \"kubernetes.io/projected/667cc0c8-9292-4cbd-a013-bd645c4b57d7-kube-api-access-xvpzr\") pod \"nova-cell0-conductor-0\" (UID: \"667cc0c8-9292-4cbd-a013-bd645c4b57d7\") " pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:49 crc kubenswrapper[4706]: I1208 19:44:49.734997 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:50 crc kubenswrapper[4706]: W1208 19:44:50.239179 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod667cc0c8_9292_4cbd_a013_bd645c4b57d7.slice/crio-7b115b628cf726dd179172fb3fc53aa13ad38a196edce435ff914783154fabf8 WatchSource:0}: Error finding container 7b115b628cf726dd179172fb3fc53aa13ad38a196edce435ff914783154fabf8: Status 404 returned error can't find the container with id 7b115b628cf726dd179172fb3fc53aa13ad38a196edce435ff914783154fabf8 Dec 08 19:44:50 crc kubenswrapper[4706]: I1208 19:44:50.247063 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 08 19:44:50 crc kubenswrapper[4706]: I1208 19:44:50.297645 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"667cc0c8-9292-4cbd-a013-bd645c4b57d7","Type":"ContainerStarted","Data":"7b115b628cf726dd179172fb3fc53aa13ad38a196edce435ff914783154fabf8"} Dec 08 19:44:51 crc kubenswrapper[4706]: I1208 19:44:51.314990 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"667cc0c8-9292-4cbd-a013-bd645c4b57d7","Type":"ContainerStarted","Data":"9a58a2cf6423b4e71f7c7136701ae6ba2182cc95948963a2357841d558fe57b0"} Dec 08 19:44:51 crc kubenswrapper[4706]: I1208 19:44:51.315870 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 08 19:44:51 crc kubenswrapper[4706]: I1208 19:44:51.337785 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.337763577 podStartE2EDuration="2.337763577s" podCreationTimestamp="2025-12-08 19:44:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:44:51.334483302 +0000 UTC m=+1373.976684305" watchObservedRunningTime="2025-12-08 19:44:51.337763577 +0000 UTC m=+1373.979964580" Dec 08 19:44:52 crc kubenswrapper[4706]: I1208 19:44:52.879372 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 08 19:44:52 crc kubenswrapper[4706]: I1208 19:44:52.879832 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 08 19:44:52 crc kubenswrapper[4706]: I1208 19:44:52.914965 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 08 19:44:52 crc kubenswrapper[4706]: I1208 19:44:52.931362 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 08 19:44:53 crc kubenswrapper[4706]: I1208 19:44:53.349479 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 08 19:44:53 crc kubenswrapper[4706]: I1208 19:44:53.349555 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 08 19:44:56 crc kubenswrapper[4706]: I1208 19:44:56.355309 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 08 19:44:56 crc kubenswrapper[4706]: I1208 19:44:56.357743 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 19:44:56 crc kubenswrapper[4706]: I1208 19:44:56.370356 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 08 19:44:59 crc kubenswrapper[4706]: I1208 19:44:59.768065 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.145030 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm"] Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.147241 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.161184 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm"] Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.198774 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.199076 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.291893 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfcdl\" (UniqueName: \"kubernetes.io/projected/652e74f8-adc5-455f-91e1-f68f9ce56533-kube-api-access-wfcdl\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.292324 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/652e74f8-adc5-455f-91e1-f68f9ce56533-secret-volume\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.292399 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/652e74f8-adc5-455f-91e1-f68f9ce56533-config-volume\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.380451 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-5vj8x"] Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.382253 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.387470 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.387796 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.394359 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5vj8x"] Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.394599 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfcdl\" (UniqueName: \"kubernetes.io/projected/652e74f8-adc5-455f-91e1-f68f9ce56533-kube-api-access-wfcdl\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.394690 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/652e74f8-adc5-455f-91e1-f68f9ce56533-secret-volume\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.394786 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/652e74f8-adc5-455f-91e1-f68f9ce56533-config-volume\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.396393 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/652e74f8-adc5-455f-91e1-f68f9ce56533-config-volume\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.405399 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/652e74f8-adc5-455f-91e1-f68f9ce56533-secret-volume\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.426017 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfcdl\" (UniqueName: \"kubernetes.io/projected/652e74f8-adc5-455f-91e1-f68f9ce56533-kube-api-access-wfcdl\") pod \"collect-profiles-29420385-hldgm\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.498078 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwxtq\" (UniqueName: \"kubernetes.io/projected/3b2d5238-1fbe-4f49-b335-566d1ebaed31-kube-api-access-mwxtq\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.498175 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.498212 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-config-data\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.498270 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-scripts\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.521000 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.600857 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-config-data\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.600937 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-scripts\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.601055 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwxtq\" (UniqueName: \"kubernetes.io/projected/3b2d5238-1fbe-4f49-b335-566d1ebaed31-kube-api-access-mwxtq\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.601119 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.605060 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.607124 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.611247 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.611646 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.621026 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-scripts\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.622109 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-config-data\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.629600 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.651424 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwxtq\" (UniqueName: \"kubernetes.io/projected/3b2d5238-1fbe-4f49-b335-566d1ebaed31-kube-api-access-mwxtq\") pod \"nova-cell0-cell-mapping-5vj8x\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.708653 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t69x\" (UniqueName: \"kubernetes.io/projected/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-kube-api-access-7t69x\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.708768 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-config-data\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.708976 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.772741 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.775161 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.778797 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.816058 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7ff6\" (UniqueName: \"kubernetes.io/projected/8105c6c2-a222-485f-a37c-48156ac7086d-kube-api-access-r7ff6\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.816128 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.816214 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t69x\" (UniqueName: \"kubernetes.io/projected/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-kube-api-access-7t69x\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.816272 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.816317 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-config-data\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.816341 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8105c6c2-a222-485f-a37c-48156ac7086d-logs\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.816368 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-config-data\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.822329 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.830299 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.851519 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-config-data\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.874044 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.919889 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t69x\" (UniqueName: \"kubernetes.io/projected/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-kube-api-access-7t69x\") pod \"nova-scheduler-0\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.931011 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7ff6\" (UniqueName: \"kubernetes.io/projected/8105c6c2-a222-485f-a37c-48156ac7086d-kube-api-access-r7ff6\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.931404 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.931461 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8105c6c2-a222-485f-a37c-48156ac7086d-logs\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.931509 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-config-data\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.933900 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8105c6c2-a222-485f-a37c-48156ac7086d-logs\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.957164 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.963935 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-config-data\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:00 crc kubenswrapper[4706]: I1208 19:45:00.998150 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7ff6\" (UniqueName: \"kubernetes.io/projected/8105c6c2-a222-485f-a37c-48156ac7086d-kube-api-access-r7ff6\") pod \"nova-api-0\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " pod="openstack/nova-api-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.027462 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.060564 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.078026 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.085567 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.089904 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d090007-7301-4b41-af03-03f4dbfdca91-logs\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.090023 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-config-data\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.090243 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xrcp\" (UniqueName: \"kubernetes.io/projected/9d090007-7301-4b41-af03-03f4dbfdca91-kube-api-access-2xrcp\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.090303 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.147462 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.158721 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.173614 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.175343 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.200176 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d090007-7301-4b41-af03-03f4dbfdca91-logs\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.200280 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-config-data\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.200432 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xrcp\" (UniqueName: \"kubernetes.io/projected/9d090007-7301-4b41-af03-03f4dbfdca91-kube-api-access-2xrcp\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.200465 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.203550 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d090007-7301-4b41-af03-03f4dbfdca91-logs\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.217327 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-config-data\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.234751 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.249790 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.262381 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xrcp\" (UniqueName: \"kubernetes.io/projected/9d090007-7301-4b41-af03-03f4dbfdca91-kube-api-access-2xrcp\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.262740 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.305780 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78cd565959-n79s4"] Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.325991 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.329756 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45pkz\" (UniqueName: \"kubernetes.io/projected/5f4bf211-db03-4068-b530-f4d58b704c59-kube-api-access-45pkz\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.330004 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.333321 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-n79s4"] Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.334420 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.435731 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.438605 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45pkz\" (UniqueName: \"kubernetes.io/projected/5f4bf211-db03-4068-b530-f4d58b704c59-kube-api-access-45pkz\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.438661 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-svc\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.438705 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.438726 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.438818 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-config\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.438878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpv25\" (UniqueName: \"kubernetes.io/projected/deed8943-263a-449c-b219-72a8c0a838dd-kube-api-access-qpv25\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.438898 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.439963 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.440117 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.444847 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.450364 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.485117 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45pkz\" (UniqueName: \"kubernetes.io/projected/5f4bf211-db03-4068-b530-f4d58b704c59-kube-api-access-45pkz\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.543000 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-config\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.543101 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpv25\" (UniqueName: \"kubernetes.io/projected/deed8943-263a-449c-b219-72a8c0a838dd-kube-api-access-qpv25\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.543128 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.543171 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.543253 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-svc\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.543342 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.546305 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.548432 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-svc\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.549621 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.551934 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-config\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.553475 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.555150 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.586916 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpv25\" (UniqueName: \"kubernetes.io/projected/deed8943-263a-449c-b219-72a8c0a838dd-kube-api-access-qpv25\") pod \"dnsmasq-dns-78cd565959-n79s4\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.706114 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:01 crc kubenswrapper[4706]: I1208 19:45:01.771284 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm"] Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.032867 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5vj8x"] Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.546618 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5vj8x" event={"ID":"3b2d5238-1fbe-4f49-b335-566d1ebaed31","Type":"ContainerStarted","Data":"47f68655d5e808bd4404524ff9da233aaf93866810ace1ed32992b1dd24e7c32"} Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.547324 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5vj8x" event={"ID":"3b2d5238-1fbe-4f49-b335-566d1ebaed31","Type":"ContainerStarted","Data":"868a90c2b231f878c742a879ee37f2c9d4ee75c4f3a7002f87bfddc778cb5752"} Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.576819 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" event={"ID":"652e74f8-adc5-455f-91e1-f68f9ce56533","Type":"ContainerStarted","Data":"78cd172ce176672df1f8a50b3ebf190e017b13f306de53cd5e52215b247761b2"} Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.576891 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" event={"ID":"652e74f8-adc5-455f-91e1-f68f9ce56533","Type":"ContainerStarted","Data":"aac7cf4e7d99770ca7a4c87de24b5071aa5c8d18ca56c4b4d979bfb33914d9b2"} Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.602032 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.621580 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-5vj8x" podStartSLOduration=2.621549321 podStartE2EDuration="2.621549321s" podCreationTimestamp="2025-12-08 19:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:02.584426561 +0000 UTC m=+1385.226627564" watchObservedRunningTime="2025-12-08 19:45:02.621549321 +0000 UTC m=+1385.263750324" Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.643647 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" podStartSLOduration=2.6436208580000002 podStartE2EDuration="2.643620858s" podCreationTimestamp="2025-12-08 19:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:02.613357645 +0000 UTC m=+1385.255558668" watchObservedRunningTime="2025-12-08 19:45:02.643620858 +0000 UTC m=+1385.285821861" Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.694054 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.765646 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.777222 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lkpfq"] Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.779317 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.783370 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.783739 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.798202 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lkpfq"] Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.912950 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-config-data\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.913190 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpmm2\" (UniqueName: \"kubernetes.io/projected/ab73a222-d68a-4716-b0bb-6750440338ad-kube-api-access-gpmm2\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.913251 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-scripts\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.915210 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:02 crc kubenswrapper[4706]: I1208 19:45:02.915447 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.000478 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-n79s4"] Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.018404 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpmm2\" (UniqueName: \"kubernetes.io/projected/ab73a222-d68a-4716-b0bb-6750440338ad-kube-api-access-gpmm2\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.018563 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-scripts\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.018684 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.018860 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-config-data\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.025379 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-config-data\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.029938 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-scripts\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.032996 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.053896 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpmm2\" (UniqueName: \"kubernetes.io/projected/ab73a222-d68a-4716-b0bb-6750440338ad-kube-api-access-gpmm2\") pod \"nova-cell1-conductor-db-sync-lkpfq\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.140996 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.494738 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.646951 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-n79s4" event={"ID":"deed8943-263a-449c-b219-72a8c0a838dd","Type":"ContainerStarted","Data":"4bd49a357b0e47975ef354cc243dc5989ab884ec4a5ae8464420afebff883e61"} Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.647010 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-n79s4" event={"ID":"deed8943-263a-449c-b219-72a8c0a838dd","Type":"ContainerStarted","Data":"e001fc1c34a7726126caf28f85d86de0d5fce222e2e91eaa895a270dfcbfc81a"} Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.661100 4706 generic.go:334] "Generic (PLEG): container finished" podID="652e74f8-adc5-455f-91e1-f68f9ce56533" containerID="78cd172ce176672df1f8a50b3ebf190e017b13f306de53cd5e52215b247761b2" exitCode=0 Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.661195 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" event={"ID":"652e74f8-adc5-455f-91e1-f68f9ce56533","Type":"ContainerDied","Data":"78cd172ce176672df1f8a50b3ebf190e017b13f306de53cd5e52215b247761b2"} Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.676868 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8105c6c2-a222-485f-a37c-48156ac7086d","Type":"ContainerStarted","Data":"9a653b098f4c55013a954f798a70555096eb8d54c88aec5872668a723ba06d6d"} Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.679423 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d090007-7301-4b41-af03-03f4dbfdca91","Type":"ContainerStarted","Data":"b9c8d994b241b640a16ae7a121114b5a9cec8ccce0db898c8aa231215ff749b7"} Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.683344 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23e4e486-b7f8-420c-8439-fd1a88e2fd4c","Type":"ContainerStarted","Data":"571cfec7d3a156a1df43fcafd1030b76c186a335b76a7126da47ec6c762caa11"} Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.684825 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5f4bf211-db03-4068-b530-f4d58b704c59","Type":"ContainerStarted","Data":"747f32acab1728b62b97f7a3f57cfbf3002cbb9a5bf364b300a1b60f0b776565"} Dec 08 19:45:03 crc kubenswrapper[4706]: I1208 19:45:03.873771 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lkpfq"] Dec 08 19:45:04 crc kubenswrapper[4706]: I1208 19:45:04.725071 4706 generic.go:334] "Generic (PLEG): container finished" podID="deed8943-263a-449c-b219-72a8c0a838dd" containerID="4bd49a357b0e47975ef354cc243dc5989ab884ec4a5ae8464420afebff883e61" exitCode=0 Dec 08 19:45:04 crc kubenswrapper[4706]: I1208 19:45:04.725189 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-n79s4" event={"ID":"deed8943-263a-449c-b219-72a8c0a838dd","Type":"ContainerDied","Data":"4bd49a357b0e47975ef354cc243dc5989ab884ec4a5ae8464420afebff883e61"} Dec 08 19:45:04 crc kubenswrapper[4706]: I1208 19:45:04.743360 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" event={"ID":"ab73a222-d68a-4716-b0bb-6750440338ad","Type":"ContainerStarted","Data":"8782ed0963e46f0dd0f6d8eb39ff25ef0919a61de7ff1869b910c1928c07a19a"} Dec 08 19:45:04 crc kubenswrapper[4706]: I1208 19:45:04.743423 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" event={"ID":"ab73a222-d68a-4716-b0bb-6750440338ad","Type":"ContainerStarted","Data":"7db6aa23a463474dc97be849be37aba10fa5138d968ea183244746d146de053d"} Dec 08 19:45:05 crc kubenswrapper[4706]: I1208 19:45:05.075410 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" podStartSLOduration=3.075360754 podStartE2EDuration="3.075360754s" podCreationTimestamp="2025-12-08 19:45:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:04.833970281 +0000 UTC m=+1387.476171284" watchObservedRunningTime="2025-12-08 19:45:05.075360754 +0000 UTC m=+1387.717561757" Dec 08 19:45:05 crc kubenswrapper[4706]: I1208 19:45:05.780940 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-n79s4" event={"ID":"deed8943-263a-449c-b219-72a8c0a838dd","Type":"ContainerStarted","Data":"97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c"} Dec 08 19:45:05 crc kubenswrapper[4706]: I1208 19:45:05.781468 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:05 crc kubenswrapper[4706]: I1208 19:45:05.826898 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78cd565959-n79s4" podStartSLOduration=4.826871193 podStartE2EDuration="4.826871193s" podCreationTimestamp="2025-12-08 19:45:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:05.819954633 +0000 UTC m=+1388.462155636" watchObservedRunningTime="2025-12-08 19:45:05.826871193 +0000 UTC m=+1388.469072196" Dec 08 19:45:06 crc kubenswrapper[4706]: I1208 19:45:06.271045 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:06 crc kubenswrapper[4706]: I1208 19:45:06.286769 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.238438 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.302988 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/652e74f8-adc5-455f-91e1-f68f9ce56533-secret-volume\") pod \"652e74f8-adc5-455f-91e1-f68f9ce56533\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.303277 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/652e74f8-adc5-455f-91e1-f68f9ce56533-config-volume\") pod \"652e74f8-adc5-455f-91e1-f68f9ce56533\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.303357 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfcdl\" (UniqueName: \"kubernetes.io/projected/652e74f8-adc5-455f-91e1-f68f9ce56533-kube-api-access-wfcdl\") pod \"652e74f8-adc5-455f-91e1-f68f9ce56533\" (UID: \"652e74f8-adc5-455f-91e1-f68f9ce56533\") " Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.305609 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/652e74f8-adc5-455f-91e1-f68f9ce56533-config-volume" (OuterVolumeSpecName: "config-volume") pod "652e74f8-adc5-455f-91e1-f68f9ce56533" (UID: "652e74f8-adc5-455f-91e1-f68f9ce56533"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.319434 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/652e74f8-adc5-455f-91e1-f68f9ce56533-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "652e74f8-adc5-455f-91e1-f68f9ce56533" (UID: "652e74f8-adc5-455f-91e1-f68f9ce56533"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.319495 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/652e74f8-adc5-455f-91e1-f68f9ce56533-kube-api-access-wfcdl" (OuterVolumeSpecName: "kube-api-access-wfcdl") pod "652e74f8-adc5-455f-91e1-f68f9ce56533" (UID: "652e74f8-adc5-455f-91e1-f68f9ce56533"). InnerVolumeSpecName "kube-api-access-wfcdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.412791 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/652e74f8-adc5-455f-91e1-f68f9ce56533-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.412853 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/652e74f8-adc5-455f-91e1-f68f9ce56533-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.412863 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfcdl\" (UniqueName: \"kubernetes.io/projected/652e74f8-adc5-455f-91e1-f68f9ce56533-kube-api-access-wfcdl\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.890779 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" event={"ID":"652e74f8-adc5-455f-91e1-f68f9ce56533","Type":"ContainerDied","Data":"aac7cf4e7d99770ca7a4c87de24b5071aa5c8d18ca56c4b4d979bfb33914d9b2"} Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.890833 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aac7cf4e7d99770ca7a4c87de24b5071aa5c8d18ca56c4b4d979bfb33914d9b2" Dec 08 19:45:07 crc kubenswrapper[4706]: I1208 19:45:07.890921 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm" Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.926173 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5f4bf211-db03-4068-b530-f4d58b704c59","Type":"ContainerStarted","Data":"7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5"} Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.926250 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5f4bf211-db03-4068-b530-f4d58b704c59" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5" gracePeriod=30 Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.931769 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8105c6c2-a222-485f-a37c-48156ac7086d","Type":"ContainerStarted","Data":"55b2e3941da25571d3b5d0e07a004b0c155e97cf535dda5db47a11a07596c243"} Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.931820 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8105c6c2-a222-485f-a37c-48156ac7086d","Type":"ContainerStarted","Data":"aedca06c0de8a38aeb91c49e1a7dee2c7655d60c402e66aee647aa8727477e99"} Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.938921 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d090007-7301-4b41-af03-03f4dbfdca91","Type":"ContainerStarted","Data":"cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675"} Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.938974 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d090007-7301-4b41-af03-03f4dbfdca91","Type":"ContainerStarted","Data":"46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c"} Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.938969 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" containerName="nova-metadata-log" containerID="cri-o://46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c" gracePeriod=30 Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.939076 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" containerName="nova-metadata-metadata" containerID="cri-o://cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675" gracePeriod=30 Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.952609 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23e4e486-b7f8-420c-8439-fd1a88e2fd4c","Type":"ContainerStarted","Data":"e3fd44f9b08e0e1b5826a88105972b9939c405758b506f43c2b8e5f28aee11ea"} Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.965581 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=4.131986343 podStartE2EDuration="9.965554599s" podCreationTimestamp="2025-12-08 19:45:00 +0000 UTC" firstStartedPulling="2025-12-08 19:45:02.908034036 +0000 UTC m=+1385.550235039" lastFinishedPulling="2025-12-08 19:45:08.741602292 +0000 UTC m=+1391.383803295" observedRunningTime="2025-12-08 19:45:09.945598193 +0000 UTC m=+1392.587799196" watchObservedRunningTime="2025-12-08 19:45:09.965554599 +0000 UTC m=+1392.607755602" Dec 08 19:45:09 crc kubenswrapper[4706]: I1208 19:45:09.989920 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.973973364 podStartE2EDuration="9.98989076s" podCreationTimestamp="2025-12-08 19:45:00 +0000 UTC" firstStartedPulling="2025-12-08 19:45:02.729205007 +0000 UTC m=+1385.371406010" lastFinishedPulling="2025-12-08 19:45:08.745122403 +0000 UTC m=+1391.387323406" observedRunningTime="2025-12-08 19:45:09.974528227 +0000 UTC m=+1392.616729250" watchObservedRunningTime="2025-12-08 19:45:09.98989076 +0000 UTC m=+1392.632091763" Dec 08 19:45:10 crc kubenswrapper[4706]: I1208 19:45:10.016941 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.878378816 podStartE2EDuration="10.01691981s" podCreationTimestamp="2025-12-08 19:45:00 +0000 UTC" firstStartedPulling="2025-12-08 19:45:02.605006534 +0000 UTC m=+1385.247207547" lastFinishedPulling="2025-12-08 19:45:08.743547538 +0000 UTC m=+1391.385748541" observedRunningTime="2025-12-08 19:45:10.003005919 +0000 UTC m=+1392.645206922" watchObservedRunningTime="2025-12-08 19:45:10.01691981 +0000 UTC m=+1392.659120803" Dec 08 19:45:10 crc kubenswrapper[4706]: I1208 19:45:10.030441 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.009537961 podStartE2EDuration="10.0304147s" podCreationTimestamp="2025-12-08 19:45:00 +0000 UTC" firstStartedPulling="2025-12-08 19:45:02.661661589 +0000 UTC m=+1385.303862592" lastFinishedPulling="2025-12-08 19:45:08.682538328 +0000 UTC m=+1391.324739331" observedRunningTime="2025-12-08 19:45:10.026733623 +0000 UTC m=+1392.668934626" watchObservedRunningTime="2025-12-08 19:45:10.0304147 +0000 UTC m=+1392.672615703" Dec 08 19:45:10 crc kubenswrapper[4706]: I1208 19:45:10.974479 4706 generic.go:334] "Generic (PLEG): container finished" podID="9d090007-7301-4b41-af03-03f4dbfdca91" containerID="46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c" exitCode=143 Dec 08 19:45:10 crc kubenswrapper[4706]: I1208 19:45:10.975740 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d090007-7301-4b41-af03-03f4dbfdca91","Type":"ContainerDied","Data":"46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c"} Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.087475 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.089362 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.151429 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.176827 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.176893 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.437556 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.437626 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.556098 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.709644 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.791039 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-bd49n"] Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.791376 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" podUID="0fe2563a-a800-40b6-84e6-60249516049f" containerName="dnsmasq-dns" containerID="cri-o://991133f9b78a74b75ac109af725526c67db3191cec10f6be63ceedc3891720fb" gracePeriod=10 Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.865979 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:45:11 crc kubenswrapper[4706]: I1208 19:45:11.866311 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="908bed13-e8ef-4234-83fd-21b3f0ed218e" containerName="kube-state-metrics" containerID="cri-o://580be45a2242147e94e765fe95e7bb0fca947f855c446213b61b694f887e318f" gracePeriod=30 Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.055680 4706 generic.go:334] "Generic (PLEG): container finished" podID="0fe2563a-a800-40b6-84e6-60249516049f" containerID="991133f9b78a74b75ac109af725526c67db3191cec10f6be63ceedc3891720fb" exitCode=0 Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.056252 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" event={"ID":"0fe2563a-a800-40b6-84e6-60249516049f","Type":"ContainerDied","Data":"991133f9b78a74b75ac109af725526c67db3191cec10f6be63ceedc3891720fb"} Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.065173 4706 generic.go:334] "Generic (PLEG): container finished" podID="908bed13-e8ef-4234-83fd-21b3f0ed218e" containerID="580be45a2242147e94e765fe95e7bb0fca947f855c446213b61b694f887e318f" exitCode=2 Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.067371 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"908bed13-e8ef-4234-83fd-21b3f0ed218e","Type":"ContainerDied","Data":"580be45a2242147e94e765fe95e7bb0fca947f855c446213b61b694f887e318f"} Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.129943 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.263609 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.263658 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.844543 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.858205 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.939529 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msfx2\" (UniqueName: \"kubernetes.io/projected/0fe2563a-a800-40b6-84e6-60249516049f-kube-api-access-msfx2\") pod \"0fe2563a-a800-40b6-84e6-60249516049f\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.939581 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-swift-storage-0\") pod \"0fe2563a-a800-40b6-84e6-60249516049f\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.939652 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-svc\") pod \"0fe2563a-a800-40b6-84e6-60249516049f\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.939967 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s94p\" (UniqueName: \"kubernetes.io/projected/908bed13-e8ef-4234-83fd-21b3f0ed218e-kube-api-access-8s94p\") pod \"908bed13-e8ef-4234-83fd-21b3f0ed218e\" (UID: \"908bed13-e8ef-4234-83fd-21b3f0ed218e\") " Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.940007 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-config\") pod \"0fe2563a-a800-40b6-84e6-60249516049f\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.940026 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-sb\") pod \"0fe2563a-a800-40b6-84e6-60249516049f\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.940059 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-nb\") pod \"0fe2563a-a800-40b6-84e6-60249516049f\" (UID: \"0fe2563a-a800-40b6-84e6-60249516049f\") " Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.955096 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fe2563a-a800-40b6-84e6-60249516049f-kube-api-access-msfx2" (OuterVolumeSpecName: "kube-api-access-msfx2") pod "0fe2563a-a800-40b6-84e6-60249516049f" (UID: "0fe2563a-a800-40b6-84e6-60249516049f"). InnerVolumeSpecName "kube-api-access-msfx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:12 crc kubenswrapper[4706]: I1208 19:45:12.965655 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/908bed13-e8ef-4234-83fd-21b3f0ed218e-kube-api-access-8s94p" (OuterVolumeSpecName: "kube-api-access-8s94p") pod "908bed13-e8ef-4234-83fd-21b3f0ed218e" (UID: "908bed13-e8ef-4234-83fd-21b3f0ed218e"). InnerVolumeSpecName "kube-api-access-8s94p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.046230 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s94p\" (UniqueName: \"kubernetes.io/projected/908bed13-e8ef-4234-83fd-21b3f0ed218e-kube-api-access-8s94p\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.046296 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msfx2\" (UniqueName: \"kubernetes.io/projected/0fe2563a-a800-40b6-84e6-60249516049f-kube-api-access-msfx2\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.126888 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-config" (OuterVolumeSpecName: "config") pod "0fe2563a-a800-40b6-84e6-60249516049f" (UID: "0fe2563a-a800-40b6-84e6-60249516049f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.148213 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.160153 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0fe2563a-a800-40b6-84e6-60249516049f" (UID: "0fe2563a-a800-40b6-84e6-60249516049f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.160591 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0fe2563a-a800-40b6-84e6-60249516049f" (UID: "0fe2563a-a800-40b6-84e6-60249516049f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.161092 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" event={"ID":"0fe2563a-a800-40b6-84e6-60249516049f","Type":"ContainerDied","Data":"d406d74d6b593282ecb796de82b926be9fbc988a4769fea75ed49644fba25c4e"} Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.161116 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-bd49n" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.161153 4706 scope.go:117] "RemoveContainer" containerID="991133f9b78a74b75ac109af725526c67db3191cec10f6be63ceedc3891720fb" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.172009 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0fe2563a-a800-40b6-84e6-60249516049f" (UID: "0fe2563a-a800-40b6-84e6-60249516049f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.181799 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.183492 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"908bed13-e8ef-4234-83fd-21b3f0ed218e","Type":"ContainerDied","Data":"d50e959f83cc5259cb909c1a186034e36ca0c871fe97b949cd978ebd975e868d"} Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.228007 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0fe2563a-a800-40b6-84e6-60249516049f" (UID: "0fe2563a-a800-40b6-84e6-60249516049f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.250572 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.250632 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.250645 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.250659 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fe2563a-a800-40b6-84e6-60249516049f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.282142 4706 scope.go:117] "RemoveContainer" containerID="8206afee5c336eaf2932dd8e51d32478ab5309323c56c29a56d09336c13fb877" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.312985 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.326991 4706 scope.go:117] "RemoveContainer" containerID="580be45a2242147e94e765fe95e7bb0fca947f855c446213b61b694f887e318f" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.332612 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.356427 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:45:13 crc kubenswrapper[4706]: E1208 19:45:13.357230 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="908bed13-e8ef-4234-83fd-21b3f0ed218e" containerName="kube-state-metrics" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.357252 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="908bed13-e8ef-4234-83fd-21b3f0ed218e" containerName="kube-state-metrics" Dec 08 19:45:13 crc kubenswrapper[4706]: E1208 19:45:13.357308 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe2563a-a800-40b6-84e6-60249516049f" containerName="dnsmasq-dns" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.357318 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe2563a-a800-40b6-84e6-60249516049f" containerName="dnsmasq-dns" Dec 08 19:45:13 crc kubenswrapper[4706]: E1208 19:45:13.357332 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe2563a-a800-40b6-84e6-60249516049f" containerName="init" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.357339 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe2563a-a800-40b6-84e6-60249516049f" containerName="init" Dec 08 19:45:13 crc kubenswrapper[4706]: E1208 19:45:13.357353 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652e74f8-adc5-455f-91e1-f68f9ce56533" containerName="collect-profiles" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.357359 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="652e74f8-adc5-455f-91e1-f68f9ce56533" containerName="collect-profiles" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.357679 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="908bed13-e8ef-4234-83fd-21b3f0ed218e" containerName="kube-state-metrics" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.357707 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="652e74f8-adc5-455f-91e1-f68f9ce56533" containerName="collect-profiles" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.357717 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fe2563a-a800-40b6-84e6-60249516049f" containerName="dnsmasq-dns" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.358763 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.363658 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.363748 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.366490 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.464295 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94zb7\" (UniqueName: \"kubernetes.io/projected/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-api-access-94zb7\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.464355 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.464411 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.464509 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.515328 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-bd49n"] Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.527054 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-bd49n"] Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.566701 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.566929 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94zb7\" (UniqueName: \"kubernetes.io/projected/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-api-access-94zb7\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.566956 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.567437 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.572382 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.572501 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.574223 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.588206 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94zb7\" (UniqueName: \"kubernetes.io/projected/aa4c62bf-4f94-4e1e-afd5-9036da150cc4-kube-api-access-94zb7\") pod \"kube-state-metrics-0\" (UID: \"aa4c62bf-4f94-4e1e-afd5-9036da150cc4\") " pod="openstack/kube-state-metrics-0" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.623919 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fe2563a-a800-40b6-84e6-60249516049f" path="/var/lib/kubelet/pods/0fe2563a-a800-40b6-84e6-60249516049f/volumes" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.624706 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="908bed13-e8ef-4234-83fd-21b3f0ed218e" path="/var/lib/kubelet/pods/908bed13-e8ef-4234-83fd-21b3f0ed218e/volumes" Dec 08 19:45:13 crc kubenswrapper[4706]: I1208 19:45:13.693838 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 19:45:14 crc kubenswrapper[4706]: W1208 19:45:14.531517 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa4c62bf_4f94_4e1e_afd5_9036da150cc4.slice/crio-76f8f10dc3fff79975c1d9205ea724301d24ec58888f8c98cfdb7e1e513d2c03 WatchSource:0}: Error finding container 76f8f10dc3fff79975c1d9205ea724301d24ec58888f8c98cfdb7e1e513d2c03: Status 404 returned error can't find the container with id 76f8f10dc3fff79975c1d9205ea724301d24ec58888f8c98cfdb7e1e513d2c03 Dec 08 19:45:14 crc kubenswrapper[4706]: I1208 19:45:14.540813 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 19:45:15 crc kubenswrapper[4706]: I1208 19:45:15.249686 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa4c62bf-4f94-4e1e-afd5-9036da150cc4","Type":"ContainerStarted","Data":"76f8f10dc3fff79975c1d9205ea724301d24ec58888f8c98cfdb7e1e513d2c03"} Dec 08 19:45:16 crc kubenswrapper[4706]: I1208 19:45:16.264914 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa4c62bf-4f94-4e1e-afd5-9036da150cc4","Type":"ContainerStarted","Data":"3759748f9650515a98f8d6653a3e6bb93d9f91d74de439fbb752d298a7b5e030"} Dec 08 19:45:16 crc kubenswrapper[4706]: I1208 19:45:16.265572 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 08 19:45:16 crc kubenswrapper[4706]: I1208 19:45:16.300794 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.92274257 podStartE2EDuration="3.300766505s" podCreationTimestamp="2025-12-08 19:45:13 +0000 UTC" firstStartedPulling="2025-12-08 19:45:14.53743412 +0000 UTC m=+1397.179635123" lastFinishedPulling="2025-12-08 19:45:14.915458055 +0000 UTC m=+1397.557659058" observedRunningTime="2025-12-08 19:45:16.289188921 +0000 UTC m=+1398.931389924" watchObservedRunningTime="2025-12-08 19:45:16.300766505 +0000 UTC m=+1398.942967508" Dec 08 19:45:16 crc kubenswrapper[4706]: I1208 19:45:16.455476 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:16 crc kubenswrapper[4706]: I1208 19:45:16.455911 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="ceilometer-central-agent" containerID="cri-o://9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef" gracePeriod=30 Dec 08 19:45:16 crc kubenswrapper[4706]: I1208 19:45:16.455999 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="sg-core" containerID="cri-o://23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940" gracePeriod=30 Dec 08 19:45:16 crc kubenswrapper[4706]: I1208 19:45:16.456027 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="ceilometer-notification-agent" containerID="cri-o://f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537" gracePeriod=30 Dec 08 19:45:16 crc kubenswrapper[4706]: I1208 19:45:16.456027 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="proxy-httpd" containerID="cri-o://68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5" gracePeriod=30 Dec 08 19:45:17 crc kubenswrapper[4706]: I1208 19:45:17.280017 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerID="68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5" exitCode=0 Dec 08 19:45:17 crc kubenswrapper[4706]: I1208 19:45:17.280412 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerID="23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940" exitCode=2 Dec 08 19:45:17 crc kubenswrapper[4706]: I1208 19:45:17.281598 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerDied","Data":"68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5"} Dec 08 19:45:17 crc kubenswrapper[4706]: I1208 19:45:17.281699 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerDied","Data":"23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940"} Dec 08 19:45:18 crc kubenswrapper[4706]: I1208 19:45:18.295339 4706 generic.go:334] "Generic (PLEG): container finished" podID="ab73a222-d68a-4716-b0bb-6750440338ad" containerID="8782ed0963e46f0dd0f6d8eb39ff25ef0919a61de7ff1869b910c1928c07a19a" exitCode=0 Dec 08 19:45:18 crc kubenswrapper[4706]: I1208 19:45:18.295428 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" event={"ID":"ab73a222-d68a-4716-b0bb-6750440338ad","Type":"ContainerDied","Data":"8782ed0963e46f0dd0f6d8eb39ff25ef0919a61de7ff1869b910c1928c07a19a"} Dec 08 19:45:18 crc kubenswrapper[4706]: I1208 19:45:18.300584 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerID="9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef" exitCode=0 Dec 08 19:45:18 crc kubenswrapper[4706]: I1208 19:45:18.300664 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerDied","Data":"9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef"} Dec 08 19:45:18 crc kubenswrapper[4706]: I1208 19:45:18.304332 4706 generic.go:334] "Generic (PLEG): container finished" podID="3b2d5238-1fbe-4f49-b335-566d1ebaed31" containerID="47f68655d5e808bd4404524ff9da233aaf93866810ace1ed32992b1dd24e7c32" exitCode=0 Dec 08 19:45:18 crc kubenswrapper[4706]: I1208 19:45:18.304388 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5vj8x" event={"ID":"3b2d5238-1fbe-4f49-b335-566d1ebaed31","Type":"ContainerDied","Data":"47f68655d5e808bd4404524ff9da233aaf93866810ace1ed32992b1dd24e7c32"} Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.095142 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.101729 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.172181 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-combined-ca-bundle\") pod \"ab73a222-d68a-4716-b0bb-6750440338ad\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.172432 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-config-data\") pod \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.172513 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-scripts\") pod \"ab73a222-d68a-4716-b0bb-6750440338ad\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.172580 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-config-data\") pod \"ab73a222-d68a-4716-b0bb-6750440338ad\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.172648 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-scripts\") pod \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.172675 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpmm2\" (UniqueName: \"kubernetes.io/projected/ab73a222-d68a-4716-b0bb-6750440338ad-kube-api-access-gpmm2\") pod \"ab73a222-d68a-4716-b0bb-6750440338ad\" (UID: \"ab73a222-d68a-4716-b0bb-6750440338ad\") " Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.172733 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-combined-ca-bundle\") pod \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.172777 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwxtq\" (UniqueName: \"kubernetes.io/projected/3b2d5238-1fbe-4f49-b335-566d1ebaed31-kube-api-access-mwxtq\") pod \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\" (UID: \"3b2d5238-1fbe-4f49-b335-566d1ebaed31\") " Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.196468 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-scripts" (OuterVolumeSpecName: "scripts") pod "3b2d5238-1fbe-4f49-b335-566d1ebaed31" (UID: "3b2d5238-1fbe-4f49-b335-566d1ebaed31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.205397 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b2d5238-1fbe-4f49-b335-566d1ebaed31-kube-api-access-mwxtq" (OuterVolumeSpecName: "kube-api-access-mwxtq") pod "3b2d5238-1fbe-4f49-b335-566d1ebaed31" (UID: "3b2d5238-1fbe-4f49-b335-566d1ebaed31"). InnerVolumeSpecName "kube-api-access-mwxtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.209456 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-scripts" (OuterVolumeSpecName: "scripts") pod "ab73a222-d68a-4716-b0bb-6750440338ad" (UID: "ab73a222-d68a-4716-b0bb-6750440338ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.232273 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab73a222-d68a-4716-b0bb-6750440338ad-kube-api-access-gpmm2" (OuterVolumeSpecName: "kube-api-access-gpmm2") pod "ab73a222-d68a-4716-b0bb-6750440338ad" (UID: "ab73a222-d68a-4716-b0bb-6750440338ad"). InnerVolumeSpecName "kube-api-access-gpmm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.271968 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-config-data" (OuterVolumeSpecName: "config-data") pod "3b2d5238-1fbe-4f49-b335-566d1ebaed31" (UID: "3b2d5238-1fbe-4f49-b335-566d1ebaed31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.272438 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b2d5238-1fbe-4f49-b335-566d1ebaed31" (UID: "3b2d5238-1fbe-4f49-b335-566d1ebaed31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.276538 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.276577 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.276591 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpmm2\" (UniqueName: \"kubernetes.io/projected/ab73a222-d68a-4716-b0bb-6750440338ad-kube-api-access-gpmm2\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.276605 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.276615 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwxtq\" (UniqueName: \"kubernetes.io/projected/3b2d5238-1fbe-4f49-b335-566d1ebaed31-kube-api-access-mwxtq\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.276625 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b2d5238-1fbe-4f49-b335-566d1ebaed31-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.285517 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab73a222-d68a-4716-b0bb-6750440338ad" (UID: "ab73a222-d68a-4716-b0bb-6750440338ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.297701 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-config-data" (OuterVolumeSpecName: "config-data") pod "ab73a222-d68a-4716-b0bb-6750440338ad" (UID: "ab73a222-d68a-4716-b0bb-6750440338ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.329309 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5vj8x" event={"ID":"3b2d5238-1fbe-4f49-b335-566d1ebaed31","Type":"ContainerDied","Data":"868a90c2b231f878c742a879ee37f2c9d4ee75c4f3a7002f87bfddc778cb5752"} Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.329379 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="868a90c2b231f878c742a879ee37f2c9d4ee75c4f3a7002f87bfddc778cb5752" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.329462 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5vj8x" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.338241 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" event={"ID":"ab73a222-d68a-4716-b0bb-6750440338ad","Type":"ContainerDied","Data":"7db6aa23a463474dc97be849be37aba10fa5138d968ea183244746d146de053d"} Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.338319 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7db6aa23a463474dc97be849be37aba10fa5138d968ea183244746d146de053d" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.338395 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lkpfq" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.390569 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.390625 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab73a222-d68a-4716-b0bb-6750440338ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.467982 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 08 19:45:20 crc kubenswrapper[4706]: E1208 19:45:20.468782 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab73a222-d68a-4716-b0bb-6750440338ad" containerName="nova-cell1-conductor-db-sync" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.468811 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab73a222-d68a-4716-b0bb-6750440338ad" containerName="nova-cell1-conductor-db-sync" Dec 08 19:45:20 crc kubenswrapper[4706]: E1208 19:45:20.468850 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b2d5238-1fbe-4f49-b335-566d1ebaed31" containerName="nova-manage" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.468860 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b2d5238-1fbe-4f49-b335-566d1ebaed31" containerName="nova-manage" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.469137 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab73a222-d68a-4716-b0bb-6750440338ad" containerName="nova-cell1-conductor-db-sync" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.469180 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b2d5238-1fbe-4f49-b335-566d1ebaed31" containerName="nova-manage" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.471446 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.474638 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.481282 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.575448 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.575830 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-log" containerID="cri-o://aedca06c0de8a38aeb91c49e1a7dee2c7655d60c402e66aee647aa8727477e99" gracePeriod=30 Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.575986 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-api" containerID="cri-o://55b2e3941da25571d3b5d0e07a004b0c155e97cf535dda5db47a11a07596c243" gracePeriod=30 Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.592855 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.593167 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="23e4e486-b7f8-420c-8439-fd1a88e2fd4c" containerName="nova-scheduler-scheduler" containerID="cri-o://e3fd44f9b08e0e1b5826a88105972b9939c405758b506f43c2b8e5f28aee11ea" gracePeriod=30 Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.596707 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/734cc0d6-b154-4980-afbe-67d4d1b9b675-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.596790 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7kkw\" (UniqueName: \"kubernetes.io/projected/734cc0d6-b154-4980-afbe-67d4d1b9b675-kube-api-access-l7kkw\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.596924 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/734cc0d6-b154-4980-afbe-67d4d1b9b675-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.699481 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/734cc0d6-b154-4980-afbe-67d4d1b9b675-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.700011 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7kkw\" (UniqueName: \"kubernetes.io/projected/734cc0d6-b154-4980-afbe-67d4d1b9b675-kube-api-access-l7kkw\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.700272 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/734cc0d6-b154-4980-afbe-67d4d1b9b675-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.714383 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/734cc0d6-b154-4980-afbe-67d4d1b9b675-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.721345 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/734cc0d6-b154-4980-afbe-67d4d1b9b675-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.723487 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7kkw\" (UniqueName: \"kubernetes.io/projected/734cc0d6-b154-4980-afbe-67d4d1b9b675-kube-api-access-l7kkw\") pod \"nova-cell1-conductor-0\" (UID: \"734cc0d6-b154-4980-afbe-67d4d1b9b675\") " pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:20 crc kubenswrapper[4706]: I1208 19:45:20.801689 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.091857 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e3fd44f9b08e0e1b5826a88105972b9939c405758b506f43c2b8e5f28aee11ea" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.094453 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e3fd44f9b08e0e1b5826a88105972b9939c405758b506f43c2b8e5f28aee11ea" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.096627 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e3fd44f9b08e0e1b5826a88105972b9939c405758b506f43c2b8e5f28aee11ea" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.096755 4706 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="23e4e486-b7f8-420c-8439-fd1a88e2fd4c" containerName="nova-scheduler-scheduler" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.268589 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.318978 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-combined-ca-bundle\") pod \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.319116 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-config-data\") pod \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.319169 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-scripts\") pod \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.319209 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-sg-core-conf-yaml\") pod \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.319389 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-run-httpd\") pod \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.319435 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-log-httpd\") pod \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.319510 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqnlb\" (UniqueName: \"kubernetes.io/projected/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-kube-api-access-tqnlb\") pod \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\" (UID: \"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91\") " Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.321049 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" (UID: "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.321796 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" (UID: "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.329562 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-kube-api-access-tqnlb" (OuterVolumeSpecName: "kube-api-access-tqnlb") pod "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" (UID: "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91"). InnerVolumeSpecName "kube-api-access-tqnlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.351329 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-scripts" (OuterVolumeSpecName: "scripts") pod "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" (UID: "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.355218 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" (UID: "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.388641 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerID="f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537" exitCode=0 Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.388766 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerDied","Data":"f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537"} Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.388811 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91","Type":"ContainerDied","Data":"17b0b6bcc14fc233c319dad0dfdcdff0012767ebde787433b0dc828c158eb113"} Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.388837 4706 scope.go:117] "RemoveContainer" containerID="68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.388845 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.402814 4706 generic.go:334] "Generic (PLEG): container finished" podID="8105c6c2-a222-485f-a37c-48156ac7086d" containerID="aedca06c0de8a38aeb91c49e1a7dee2c7655d60c402e66aee647aa8727477e99" exitCode=143 Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.402880 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8105c6c2-a222-485f-a37c-48156ac7086d","Type":"ContainerDied","Data":"aedca06c0de8a38aeb91c49e1a7dee2c7655d60c402e66aee647aa8727477e99"} Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.427201 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.427248 4706 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.427281 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.427292 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.427302 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqnlb\" (UniqueName: \"kubernetes.io/projected/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-kube-api-access-tqnlb\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.459367 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.462305 4706 scope.go:117] "RemoveContainer" containerID="23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.509789 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-config-data" (OuterVolumeSpecName: "config-data") pod "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" (UID: "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.514525 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" (UID: "bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.527442 4706 scope.go:117] "RemoveContainer" containerID="f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.529416 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.529446 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.585940 4706 scope.go:117] "RemoveContainer" containerID="9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.663743 4706 scope.go:117] "RemoveContainer" containerID="68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5" Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.664375 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5\": container with ID starting with 68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5 not found: ID does not exist" containerID="68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.664433 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5"} err="failed to get container status \"68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5\": rpc error: code = NotFound desc = could not find container \"68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5\": container with ID starting with 68164359c4872d6e4aa9b8991530c81b407c9392ca2202d0d6a2f4ad9ea899e5 not found: ID does not exist" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.664468 4706 scope.go:117] "RemoveContainer" containerID="23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940" Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.664853 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940\": container with ID starting with 23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940 not found: ID does not exist" containerID="23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.664959 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940"} err="failed to get container status \"23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940\": rpc error: code = NotFound desc = could not find container \"23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940\": container with ID starting with 23047d34d4bf7f94e718423a4f841a261bc6da47ebcb2a8ff2a99932e51ef940 not found: ID does not exist" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.665059 4706 scope.go:117] "RemoveContainer" containerID="f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537" Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.665590 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537\": container with ID starting with f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537 not found: ID does not exist" containerID="f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.665628 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537"} err="failed to get container status \"f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537\": rpc error: code = NotFound desc = could not find container \"f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537\": container with ID starting with f9cd44558136d629fd2e266a5e4eba79576de7bb4c6013cd83352fc3b7b35537 not found: ID does not exist" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.665652 4706 scope.go:117] "RemoveContainer" containerID="9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef" Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.665976 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef\": container with ID starting with 9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef not found: ID does not exist" containerID="9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.666012 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef"} err="failed to get container status \"9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef\": rpc error: code = NotFound desc = could not find container \"9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef\": container with ID starting with 9041c8acad4a2811cb2865f989302ec5aedb44fa2b839ea45cabc996c09992ef not found: ID does not exist" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.726086 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.743350 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.794457 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.799802 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="ceilometer-central-agent" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.799865 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="ceilometer-central-agent" Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.799884 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="ceilometer-notification-agent" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.799895 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="ceilometer-notification-agent" Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.799923 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="proxy-httpd" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.799932 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="proxy-httpd" Dec 08 19:45:21 crc kubenswrapper[4706]: E1208 19:45:21.799967 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="sg-core" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.799976 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="sg-core" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.800845 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="proxy-httpd" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.800884 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="ceilometer-notification-agent" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.800940 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="ceilometer-central-agent" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.800978 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" containerName="sg-core" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.819513 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.824008 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.824298 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.824448 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.841847 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.942898 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.943084 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-run-httpd\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.943123 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.943139 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.943203 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-log-httpd\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.943229 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-config-data\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.943248 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwf8k\" (UniqueName: \"kubernetes.io/projected/b6802220-f113-4842-b5d1-d7506f332423-kube-api-access-rwf8k\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:21 crc kubenswrapper[4706]: I1208 19:45:21.943305 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-scripts\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.045795 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-run-httpd\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.045901 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.045944 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.046028 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-log-httpd\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.046054 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-config-data\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.046075 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwf8k\" (UniqueName: \"kubernetes.io/projected/b6802220-f113-4842-b5d1-d7506f332423-kube-api-access-rwf8k\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.046113 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-scripts\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.046132 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.047299 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-run-httpd\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.047312 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-log-httpd\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.052656 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.053208 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.055691 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-scripts\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.066997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwf8k\" (UniqueName: \"kubernetes.io/projected/b6802220-f113-4842-b5d1-d7506f332423-kube-api-access-rwf8k\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.069734 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-config-data\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.071475 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.215501 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.442413 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"734cc0d6-b154-4980-afbe-67d4d1b9b675","Type":"ContainerStarted","Data":"1b3299ad5c3f50ed4e18141e1ec8800636f38cd4c6cfc97dd75cd34b69674680"} Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.442835 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"734cc0d6-b154-4980-afbe-67d4d1b9b675","Type":"ContainerStarted","Data":"61a74f9b52f7bcb9097658660ba85bdcbd05fe5818bba4b51570c141bc109270"} Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.444570 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.447443 4706 generic.go:334] "Generic (PLEG): container finished" podID="23e4e486-b7f8-420c-8439-fd1a88e2fd4c" containerID="e3fd44f9b08e0e1b5826a88105972b9939c405758b506f43c2b8e5f28aee11ea" exitCode=0 Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.447490 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23e4e486-b7f8-420c-8439-fd1a88e2fd4c","Type":"ContainerDied","Data":"e3fd44f9b08e0e1b5826a88105972b9939c405758b506f43c2b8e5f28aee11ea"} Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.483503 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.483473424 podStartE2EDuration="2.483473424s" podCreationTimestamp="2025-12-08 19:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:22.472683413 +0000 UTC m=+1405.114884416" watchObservedRunningTime="2025-12-08 19:45:22.483473424 +0000 UTC m=+1405.125674427" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.540663 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.670419 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-combined-ca-bundle\") pod \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.670758 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t69x\" (UniqueName: \"kubernetes.io/projected/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-kube-api-access-7t69x\") pod \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.670987 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-config-data\") pod \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\" (UID: \"23e4e486-b7f8-420c-8439-fd1a88e2fd4c\") " Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.683758 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-kube-api-access-7t69x" (OuterVolumeSpecName: "kube-api-access-7t69x") pod "23e4e486-b7f8-420c-8439-fd1a88e2fd4c" (UID: "23e4e486-b7f8-420c-8439-fd1a88e2fd4c"). InnerVolumeSpecName "kube-api-access-7t69x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.778904 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t69x\" (UniqueName: \"kubernetes.io/projected/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-kube-api-access-7t69x\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.785589 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23e4e486-b7f8-420c-8439-fd1a88e2fd4c" (UID: "23e4e486-b7f8-420c-8439-fd1a88e2fd4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.805753 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-config-data" (OuterVolumeSpecName: "config-data") pod "23e4e486-b7f8-420c-8439-fd1a88e2fd4c" (UID: "23e4e486-b7f8-420c-8439-fd1a88e2fd4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.882427 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.882842 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e4e486-b7f8-420c-8439-fd1a88e2fd4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:22 crc kubenswrapper[4706]: W1208 19:45:22.946215 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6802220_f113_4842_b5d1_d7506f332423.slice/crio-0294837cae3888419b2499a1d081632c0c482c7cc48ce4d6b273f84ae7e0f5da WatchSource:0}: Error finding container 0294837cae3888419b2499a1d081632c0c482c7cc48ce4d6b273f84ae7e0f5da: Status 404 returned error can't find the container with id 0294837cae3888419b2499a1d081632c0c482c7cc48ce4d6b273f84ae7e0f5da Dec 08 19:45:22 crc kubenswrapper[4706]: I1208 19:45:22.957678 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.460490 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerStarted","Data":"0294837cae3888419b2499a1d081632c0c482c7cc48ce4d6b273f84ae7e0f5da"} Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.463430 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23e4e486-b7f8-420c-8439-fd1a88e2fd4c","Type":"ContainerDied","Data":"571cfec7d3a156a1df43fcafd1030b76c186a335b76a7126da47ec6c762caa11"} Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.463497 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.463536 4706 scope.go:117] "RemoveContainer" containerID="e3fd44f9b08e0e1b5826a88105972b9939c405758b506f43c2b8e5f28aee11ea" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.515412 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.533554 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.549550 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:23 crc kubenswrapper[4706]: E1208 19:45:23.550316 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e4e486-b7f8-420c-8439-fd1a88e2fd4c" containerName="nova-scheduler-scheduler" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.550342 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e4e486-b7f8-420c-8439-fd1a88e2fd4c" containerName="nova-scheduler-scheduler" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.550683 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="23e4e486-b7f8-420c-8439-fd1a88e2fd4c" containerName="nova-scheduler-scheduler" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.551999 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.558932 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.577396 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.602589 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-config-data\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.602874 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8867d\" (UniqueName: \"kubernetes.io/projected/713c3abb-338f-4a61-a52d-c3447a7dceec-kube-api-access-8867d\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.603067 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.623128 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23e4e486-b7f8-420c-8439-fd1a88e2fd4c" path="/var/lib/kubelet/pods/23e4e486-b7f8-420c-8439-fd1a88e2fd4c/volumes" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.627845 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91" path="/var/lib/kubelet/pods/bf2d01aa-0c7c-4ea2-ae58-2b25ae026a91/volumes" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.707665 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-config-data\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.708293 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8867d\" (UniqueName: \"kubernetes.io/projected/713c3abb-338f-4a61-a52d-c3447a7dceec-kube-api-access-8867d\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.708512 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.724085 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-config-data\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.724487 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.729903 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.735045 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8867d\" (UniqueName: \"kubernetes.io/projected/713c3abb-338f-4a61-a52d-c3447a7dceec-kube-api-access-8867d\") pod \"nova-scheduler-0\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " pod="openstack/nova-scheduler-0" Dec 08 19:45:23 crc kubenswrapper[4706]: I1208 19:45:23.888568 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:45:24 crc kubenswrapper[4706]: I1208 19:45:24.499352 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerStarted","Data":"950a12558264e5ceb64430e286b92d379cdd7a7add955288b0131f1dc80d2c55"} Dec 08 19:45:24 crc kubenswrapper[4706]: I1208 19:45:24.502118 4706 generic.go:334] "Generic (PLEG): container finished" podID="8105c6c2-a222-485f-a37c-48156ac7086d" containerID="55b2e3941da25571d3b5d0e07a004b0c155e97cf535dda5db47a11a07596c243" exitCode=0 Dec 08 19:45:24 crc kubenswrapper[4706]: I1208 19:45:24.502189 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8105c6c2-a222-485f-a37c-48156ac7086d","Type":"ContainerDied","Data":"55b2e3941da25571d3b5d0e07a004b0c155e97cf535dda5db47a11a07596c243"} Dec 08 19:45:24 crc kubenswrapper[4706]: I1208 19:45:24.509546 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.084684 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.159220 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-config-data\") pod \"8105c6c2-a222-485f-a37c-48156ac7086d\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.159630 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7ff6\" (UniqueName: \"kubernetes.io/projected/8105c6c2-a222-485f-a37c-48156ac7086d-kube-api-access-r7ff6\") pod \"8105c6c2-a222-485f-a37c-48156ac7086d\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.159799 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8105c6c2-a222-485f-a37c-48156ac7086d-logs\") pod \"8105c6c2-a222-485f-a37c-48156ac7086d\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.160081 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-combined-ca-bundle\") pod \"8105c6c2-a222-485f-a37c-48156ac7086d\" (UID: \"8105c6c2-a222-485f-a37c-48156ac7086d\") " Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.170349 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8105c6c2-a222-485f-a37c-48156ac7086d-logs" (OuterVolumeSpecName: "logs") pod "8105c6c2-a222-485f-a37c-48156ac7086d" (UID: "8105c6c2-a222-485f-a37c-48156ac7086d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.187595 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8105c6c2-a222-485f-a37c-48156ac7086d-kube-api-access-r7ff6" (OuterVolumeSpecName: "kube-api-access-r7ff6") pod "8105c6c2-a222-485f-a37c-48156ac7086d" (UID: "8105c6c2-a222-485f-a37c-48156ac7086d"). InnerVolumeSpecName "kube-api-access-r7ff6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.214596 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-config-data" (OuterVolumeSpecName: "config-data") pod "8105c6c2-a222-485f-a37c-48156ac7086d" (UID: "8105c6c2-a222-485f-a37c-48156ac7086d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.230381 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8105c6c2-a222-485f-a37c-48156ac7086d" (UID: "8105c6c2-a222-485f-a37c-48156ac7086d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.265864 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8105c6c2-a222-485f-a37c-48156ac7086d-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.266220 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.266384 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105c6c2-a222-485f-a37c-48156ac7086d-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.266476 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7ff6\" (UniqueName: \"kubernetes.io/projected/8105c6c2-a222-485f-a37c-48156ac7086d-kube-api-access-r7ff6\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.526655 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"713c3abb-338f-4a61-a52d-c3447a7dceec","Type":"ContainerStarted","Data":"140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc"} Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.526766 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"713c3abb-338f-4a61-a52d-c3447a7dceec","Type":"ContainerStarted","Data":"305bdb21bc95d315149718a962ea761d3710aaaac399187a2460bae2af35d554"} Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.529434 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerStarted","Data":"2f69d06b4582a22923544e57a10bef799b8c817f7847d022667b078daf68f25d"} Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.538714 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.538734 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8105c6c2-a222-485f-a37c-48156ac7086d","Type":"ContainerDied","Data":"9a653b098f4c55013a954f798a70555096eb8d54c88aec5872668a723ba06d6d"} Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.539259 4706 scope.go:117] "RemoveContainer" containerID="55b2e3941da25571d3b5d0e07a004b0c155e97cf535dda5db47a11a07596c243" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.563800 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.563766819 podStartE2EDuration="2.563766819s" podCreationTimestamp="2025-12-08 19:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:25.55132515 +0000 UTC m=+1408.193526163" watchObservedRunningTime="2025-12-08 19:45:25.563766819 +0000 UTC m=+1408.205967822" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.587822 4706 scope.go:117] "RemoveContainer" containerID="aedca06c0de8a38aeb91c49e1a7dee2c7655d60c402e66aee647aa8727477e99" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.645820 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.659514 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.671927 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:25 crc kubenswrapper[4706]: E1208 19:45:25.672585 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-log" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.672604 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-log" Dec 08 19:45:25 crc kubenswrapper[4706]: E1208 19:45:25.672648 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-api" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.672655 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-api" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.672890 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-log" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.672914 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" containerName="nova-api-api" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.674362 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.682016 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.685435 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.787392 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.787439 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-config-data\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.787574 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnphb\" (UniqueName: \"kubernetes.io/projected/928552a4-1547-43ac-afe5-645619776c5e-kube-api-access-fnphb\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.787629 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/928552a4-1547-43ac-afe5-645619776c5e-logs\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.892390 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.892439 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-config-data\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.892560 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnphb\" (UniqueName: \"kubernetes.io/projected/928552a4-1547-43ac-afe5-645619776c5e-kube-api-access-fnphb\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.892611 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/928552a4-1547-43ac-afe5-645619776c5e-logs\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.898569 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.899659 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-config-data\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.900663 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/928552a4-1547-43ac-afe5-645619776c5e-logs\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:25 crc kubenswrapper[4706]: I1208 19:45:25.964052 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnphb\" (UniqueName: \"kubernetes.io/projected/928552a4-1547-43ac-afe5-645619776c5e-kube-api-access-fnphb\") pod \"nova-api-0\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " pod="openstack/nova-api-0" Dec 08 19:45:26 crc kubenswrapper[4706]: I1208 19:45:26.017074 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:26 crc kubenswrapper[4706]: I1208 19:45:26.556987 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerStarted","Data":"1d1ae3a8cdabebc3097b78f8d921797c22ef2a0442ee9adcfca703de4e2b2d73"} Dec 08 19:45:26 crc kubenswrapper[4706]: I1208 19:45:26.839198 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:27 crc kubenswrapper[4706]: I1208 19:45:27.647888 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8105c6c2-a222-485f-a37c-48156ac7086d" path="/var/lib/kubelet/pods/8105c6c2-a222-485f-a37c-48156ac7086d/volumes" Dec 08 19:45:27 crc kubenswrapper[4706]: I1208 19:45:27.649277 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"928552a4-1547-43ac-afe5-645619776c5e","Type":"ContainerStarted","Data":"0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5"} Dec 08 19:45:27 crc kubenswrapper[4706]: I1208 19:45:27.649307 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"928552a4-1547-43ac-afe5-645619776c5e","Type":"ContainerStarted","Data":"36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267"} Dec 08 19:45:27 crc kubenswrapper[4706]: I1208 19:45:27.649320 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"928552a4-1547-43ac-afe5-645619776c5e","Type":"ContainerStarted","Data":"4f799c7e4ddde50714733fd825c8d5bf65d6128301908570036f07abfb252c26"} Dec 08 19:45:27 crc kubenswrapper[4706]: I1208 19:45:27.649330 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerStarted","Data":"49edd1079225936be761ace2702adc57714750e50e791a91da26884f0f914e17"} Dec 08 19:45:27 crc kubenswrapper[4706]: I1208 19:45:27.734753 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.734720162 podStartE2EDuration="2.734720162s" podCreationTimestamp="2025-12-08 19:45:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:27.72080049 +0000 UTC m=+1410.363001513" watchObservedRunningTime="2025-12-08 19:45:27.734720162 +0000 UTC m=+1410.376921165" Dec 08 19:45:27 crc kubenswrapper[4706]: I1208 19:45:27.760317 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.372841888 podStartE2EDuration="6.760250969s" podCreationTimestamp="2025-12-08 19:45:21 +0000 UTC" firstStartedPulling="2025-12-08 19:45:22.950025151 +0000 UTC m=+1405.592226154" lastFinishedPulling="2025-12-08 19:45:27.337434232 +0000 UTC m=+1409.979635235" observedRunningTime="2025-12-08 19:45:27.75023928 +0000 UTC m=+1410.392440283" watchObservedRunningTime="2025-12-08 19:45:27.760250969 +0000 UTC m=+1410.402451972" Dec 08 19:45:28 crc kubenswrapper[4706]: I1208 19:45:28.650001 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 19:45:28 crc kubenswrapper[4706]: I1208 19:45:28.888821 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 08 19:45:30 crc kubenswrapper[4706]: I1208 19:45:30.850627 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 08 19:45:33 crc kubenswrapper[4706]: I1208 19:45:33.888898 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 08 19:45:33 crc kubenswrapper[4706]: I1208 19:45:33.924957 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 08 19:45:34 crc kubenswrapper[4706]: I1208 19:45:34.756052 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 08 19:45:35 crc kubenswrapper[4706]: I1208 19:45:35.835607 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:45:35 crc kubenswrapper[4706]: I1208 19:45:35.835684 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:45:36 crc kubenswrapper[4706]: I1208 19:45:36.019212 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 19:45:36 crc kubenswrapper[4706]: I1208 19:45:36.019306 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 19:45:37 crc kubenswrapper[4706]: I1208 19:45:37.101603 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 08 19:45:37 crc kubenswrapper[4706]: I1208 19:45:37.101668 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.653604 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.665062 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.729765 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xrcp\" (UniqueName: \"kubernetes.io/projected/9d090007-7301-4b41-af03-03f4dbfdca91-kube-api-access-2xrcp\") pod \"9d090007-7301-4b41-af03-03f4dbfdca91\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.729875 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-config-data\") pod \"5f4bf211-db03-4068-b530-f4d58b704c59\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.729951 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45pkz\" (UniqueName: \"kubernetes.io/projected/5f4bf211-db03-4068-b530-f4d58b704c59-kube-api-access-45pkz\") pod \"5f4bf211-db03-4068-b530-f4d58b704c59\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.730112 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-config-data\") pod \"9d090007-7301-4b41-af03-03f4dbfdca91\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.730281 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-combined-ca-bundle\") pod \"9d090007-7301-4b41-af03-03f4dbfdca91\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.730350 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-combined-ca-bundle\") pod \"5f4bf211-db03-4068-b530-f4d58b704c59\" (UID: \"5f4bf211-db03-4068-b530-f4d58b704c59\") " Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.730386 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d090007-7301-4b41-af03-03f4dbfdca91-logs\") pod \"9d090007-7301-4b41-af03-03f4dbfdca91\" (UID: \"9d090007-7301-4b41-af03-03f4dbfdca91\") " Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.731600 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d090007-7301-4b41-af03-03f4dbfdca91-logs" (OuterVolumeSpecName: "logs") pod "9d090007-7301-4b41-af03-03f4dbfdca91" (UID: "9d090007-7301-4b41-af03-03f4dbfdca91"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.746731 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4bf211-db03-4068-b530-f4d58b704c59-kube-api-access-45pkz" (OuterVolumeSpecName: "kube-api-access-45pkz") pod "5f4bf211-db03-4068-b530-f4d58b704c59" (UID: "5f4bf211-db03-4068-b530-f4d58b704c59"). InnerVolumeSpecName "kube-api-access-45pkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.748284 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d090007-7301-4b41-af03-03f4dbfdca91-kube-api-access-2xrcp" (OuterVolumeSpecName: "kube-api-access-2xrcp") pod "9d090007-7301-4b41-af03-03f4dbfdca91" (UID: "9d090007-7301-4b41-af03-03f4dbfdca91"). InnerVolumeSpecName "kube-api-access-2xrcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.772441 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-config-data" (OuterVolumeSpecName: "config-data") pod "5f4bf211-db03-4068-b530-f4d58b704c59" (UID: "5f4bf211-db03-4068-b530-f4d58b704c59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.773546 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f4bf211-db03-4068-b530-f4d58b704c59" (UID: "5f4bf211-db03-4068-b530-f4d58b704c59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.775889 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-config-data" (OuterVolumeSpecName: "config-data") pod "9d090007-7301-4b41-af03-03f4dbfdca91" (UID: "9d090007-7301-4b41-af03-03f4dbfdca91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.778293 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d090007-7301-4b41-af03-03f4dbfdca91" (UID: "9d090007-7301-4b41-af03-03f4dbfdca91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.806936 4706 generic.go:334] "Generic (PLEG): container finished" podID="9d090007-7301-4b41-af03-03f4dbfdca91" containerID="cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675" exitCode=137 Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.807463 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d090007-7301-4b41-af03-03f4dbfdca91","Type":"ContainerDied","Data":"cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675"} Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.807507 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d090007-7301-4b41-af03-03f4dbfdca91","Type":"ContainerDied","Data":"b9c8d994b241b640a16ae7a121114b5a9cec8ccce0db898c8aa231215ff749b7"} Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.807528 4706 scope.go:117] "RemoveContainer" containerID="cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.807694 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.813117 4706 generic.go:334] "Generic (PLEG): container finished" podID="5f4bf211-db03-4068-b530-f4d58b704c59" containerID="7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5" exitCode=137 Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.813195 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.813174 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5f4bf211-db03-4068-b530-f4d58b704c59","Type":"ContainerDied","Data":"7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5"} Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.813392 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5f4bf211-db03-4068-b530-f4d58b704c59","Type":"ContainerDied","Data":"747f32acab1728b62b97f7a3f57cfbf3002cbb9a5bf364b300a1b60f0b776565"} Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.832195 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xrcp\" (UniqueName: \"kubernetes.io/projected/9d090007-7301-4b41-af03-03f4dbfdca91-kube-api-access-2xrcp\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.832244 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.832259 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45pkz\" (UniqueName: \"kubernetes.io/projected/5f4bf211-db03-4068-b530-f4d58b704c59-kube-api-access-45pkz\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.832292 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.832305 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d090007-7301-4b41-af03-03f4dbfdca91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.832316 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4bf211-db03-4068-b530-f4d58b704c59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.832334 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d090007-7301-4b41-af03-03f4dbfdca91-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.849092 4706 scope.go:117] "RemoveContainer" containerID="46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.854052 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.876377 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.893310 4706 scope.go:117] "RemoveContainer" containerID="cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675" Dec 08 19:45:40 crc kubenswrapper[4706]: E1208 19:45:40.894734 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675\": container with ID starting with cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675 not found: ID does not exist" containerID="cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.894798 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675"} err="failed to get container status \"cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675\": rpc error: code = NotFound desc = could not find container \"cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675\": container with ID starting with cf288b43e718767177216698b5b1a71b4cebc34a3d343c288ddf7bf399f01675 not found: ID does not exist" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.894834 4706 scope.go:117] "RemoveContainer" containerID="46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c" Dec 08 19:45:40 crc kubenswrapper[4706]: E1208 19:45:40.898911 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c\": container with ID starting with 46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c not found: ID does not exist" containerID="46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.898972 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c"} err="failed to get container status \"46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c\": rpc error: code = NotFound desc = could not find container \"46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c\": container with ID starting with 46bcee2f39576fa3865e647d8dd422621f29e9200d60f3b6ac5cf764149aa02c not found: ID does not exist" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.899014 4706 scope.go:117] "RemoveContainer" containerID="7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.904489 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.925841 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.937735 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:40 crc kubenswrapper[4706]: E1208 19:45:40.938311 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" containerName="nova-metadata-log" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.938329 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" containerName="nova-metadata-log" Dec 08 19:45:40 crc kubenswrapper[4706]: E1208 19:45:40.938365 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4bf211-db03-4068-b530-f4d58b704c59" containerName="nova-cell1-novncproxy-novncproxy" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.938375 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4bf211-db03-4068-b530-f4d58b704c59" containerName="nova-cell1-novncproxy-novncproxy" Dec 08 19:45:40 crc kubenswrapper[4706]: E1208 19:45:40.938392 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" containerName="nova-metadata-metadata" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.938399 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" containerName="nova-metadata-metadata" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.938631 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4bf211-db03-4068-b530-f4d58b704c59" containerName="nova-cell1-novncproxy-novncproxy" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.938664 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" containerName="nova-metadata-metadata" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.938679 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" containerName="nova-metadata-log" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.940625 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.943423 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.942955 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.949938 4706 scope.go:117] "RemoveContainer" containerID="7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5" Dec 08 19:45:40 crc kubenswrapper[4706]: E1208 19:45:40.950902 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5\": container with ID starting with 7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5 not found: ID does not exist" containerID="7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.950940 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5"} err="failed to get container status \"7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5\": rpc error: code = NotFound desc = could not find container \"7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5\": container with ID starting with 7246bd87fcd63e280ebc148189d0751060be32b676d7629e6a6a0301698f3fc5 not found: ID does not exist" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.956794 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.977323 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.979185 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.982629 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.982910 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 08 19:45:40 crc kubenswrapper[4706]: I1208 19:45:40.984590 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.000143 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.037719 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.038971 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.039210 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.039796 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.039855 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s9vb\" (UniqueName: \"kubernetes.io/projected/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-kube-api-access-4s9vb\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.142758 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.142855 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.142885 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s9vb\" (UniqueName: \"kubernetes.io/projected/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-kube-api-access-4s9vb\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.142924 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmh7w\" (UniqueName: \"kubernetes.io/projected/a77727be-8bf5-4c35-827b-ce13e018eaee-kube-api-access-xmh7w\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.142961 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.142993 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.143017 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a77727be-8bf5-4c35-827b-ce13e018eaee-logs\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.143077 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-config-data\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.143177 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.143208 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.148805 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.149017 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.149475 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.150219 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.162605 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s9vb\" (UniqueName: \"kubernetes.io/projected/8d07dfb3-a2ed-4e66-80a3-3288ff3f8669-kube-api-access-4s9vb\") pod \"nova-cell1-novncproxy-0\" (UID: \"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.245933 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.246452 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmh7w\" (UniqueName: \"kubernetes.io/projected/a77727be-8bf5-4c35-827b-ce13e018eaee-kube-api-access-xmh7w\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.246806 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.246938 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a77727be-8bf5-4c35-827b-ce13e018eaee-logs\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.247220 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-config-data\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.247525 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a77727be-8bf5-4c35-827b-ce13e018eaee-logs\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.250793 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.251023 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.252017 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-config-data\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.267707 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmh7w\" (UniqueName: \"kubernetes.io/projected/a77727be-8bf5-4c35-827b-ce13e018eaee-kube-api-access-xmh7w\") pod \"nova-metadata-0\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.269802 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.299596 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.651346 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f4bf211-db03-4068-b530-f4d58b704c59" path="/var/lib/kubelet/pods/5f4bf211-db03-4068-b530-f4d58b704c59/volumes" Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.652546 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d090007-7301-4b41-af03-03f4dbfdca91" path="/var/lib/kubelet/pods/9d090007-7301-4b41-af03-03f4dbfdca91/volumes" Dec 08 19:45:41 crc kubenswrapper[4706]: W1208 19:45:41.758844 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda77727be_8bf5_4c35_827b_ce13e018eaee.slice/crio-47e8fb021e3e9191fd0d8bc4c906964991b00d68c8582c3bbe4fcef4487abd37 WatchSource:0}: Error finding container 47e8fb021e3e9191fd0d8bc4c906964991b00d68c8582c3bbe4fcef4487abd37: Status 404 returned error can't find the container with id 47e8fb021e3e9191fd0d8bc4c906964991b00d68c8582c3bbe4fcef4487abd37 Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.761594 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.827226 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a77727be-8bf5-4c35-827b-ce13e018eaee","Type":"ContainerStarted","Data":"47e8fb021e3e9191fd0d8bc4c906964991b00d68c8582c3bbe4fcef4487abd37"} Dec 08 19:45:41 crc kubenswrapper[4706]: W1208 19:45:41.885418 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d07dfb3_a2ed_4e66_80a3_3288ff3f8669.slice/crio-0f9a4bf8c79c387f027c01b770db6b5de2c718dfcaa6aac5f646f234ed021bff WatchSource:0}: Error finding container 0f9a4bf8c79c387f027c01b770db6b5de2c718dfcaa6aac5f646f234ed021bff: Status 404 returned error can't find the container with id 0f9a4bf8c79c387f027c01b770db6b5de2c718dfcaa6aac5f646f234ed021bff Dec 08 19:45:41 crc kubenswrapper[4706]: I1208 19:45:41.891990 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 19:45:42 crc kubenswrapper[4706]: I1208 19:45:42.841981 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a77727be-8bf5-4c35-827b-ce13e018eaee","Type":"ContainerStarted","Data":"20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0"} Dec 08 19:45:42 crc kubenswrapper[4706]: I1208 19:45:42.842435 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a77727be-8bf5-4c35-827b-ce13e018eaee","Type":"ContainerStarted","Data":"a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180"} Dec 08 19:45:42 crc kubenswrapper[4706]: I1208 19:45:42.843792 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669","Type":"ContainerStarted","Data":"67ffe8483e02d379f9f368ca3a442810018aae6dfab721bb3fef16964b8b6fb7"} Dec 08 19:45:42 crc kubenswrapper[4706]: I1208 19:45:42.843823 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8d07dfb3-a2ed-4e66-80a3-3288ff3f8669","Type":"ContainerStarted","Data":"0f9a4bf8c79c387f027c01b770db6b5de2c718dfcaa6aac5f646f234ed021bff"} Dec 08 19:45:42 crc kubenswrapper[4706]: I1208 19:45:42.879464 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.87943154 podStartE2EDuration="2.87943154s" podCreationTimestamp="2025-12-08 19:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:42.862249845 +0000 UTC m=+1425.504450868" watchObservedRunningTime="2025-12-08 19:45:42.87943154 +0000 UTC m=+1425.521632533" Dec 08 19:45:42 crc kubenswrapper[4706]: I1208 19:45:42.906426 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.906395388 podStartE2EDuration="2.906395388s" podCreationTimestamp="2025-12-08 19:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:42.890444158 +0000 UTC m=+1425.532645181" watchObservedRunningTime="2025-12-08 19:45:42.906395388 +0000 UTC m=+1425.548596391" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.022106 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.023653 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.026473 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.026996 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.270472 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.271469 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.300028 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.894327 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 08 19:45:46 crc kubenswrapper[4706]: I1208 19:45:46.899067 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.132506 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-4bzh4"] Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.134950 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.170760 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-4bzh4"] Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.295862 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.296023 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.296135 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.296369 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.296464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrthz\" (UniqueName: \"kubernetes.io/projected/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-kube-api-access-mrthz\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.296591 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.399009 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.399078 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrthz\" (UniqueName: \"kubernetes.io/projected/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-kube-api-access-mrthz\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.399115 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.399167 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.399203 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.399247 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.400201 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.400776 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.403026 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.403054 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.403183 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.435498 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrthz\" (UniqueName: \"kubernetes.io/projected/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-kube-api-access-mrthz\") pod \"dnsmasq-dns-5fd9b586ff-4bzh4\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.467429 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:47 crc kubenswrapper[4706]: I1208 19:45:47.991982 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-4bzh4"] Dec 08 19:45:48 crc kubenswrapper[4706]: I1208 19:45:48.946245 4706 generic.go:334] "Generic (PLEG): container finished" podID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" containerID="4f1206e835725090ac898784cb6a41086707eb82045d9a4b8d17de88f475ab04" exitCode=0 Dec 08 19:45:48 crc kubenswrapper[4706]: I1208 19:45:48.950109 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" event={"ID":"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa","Type":"ContainerDied","Data":"4f1206e835725090ac898784cb6a41086707eb82045d9a4b8d17de88f475ab04"} Dec 08 19:45:48 crc kubenswrapper[4706]: I1208 19:45:48.950157 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" event={"ID":"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa","Type":"ContainerStarted","Data":"fe888104256b4b207cc2e4846ac8485159f732418062e2aee9be7e4404178404"} Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.603076 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.604342 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="proxy-httpd" containerID="cri-o://49edd1079225936be761ace2702adc57714750e50e791a91da26884f0f914e17" gracePeriod=30 Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.604372 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="ceilometer-notification-agent" containerID="cri-o://2f69d06b4582a22923544e57a10bef799b8c817f7847d022667b078daf68f25d" gracePeriod=30 Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.604209 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="ceilometer-central-agent" containerID="cri-o://950a12558264e5ceb64430e286b92d379cdd7a7add955288b0131f1dc80d2c55" gracePeriod=30 Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.604688 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="sg-core" containerID="cri-o://1d1ae3a8cdabebc3097b78f8d921797c22ef2a0442ee9adcfca703de4e2b2d73" gracePeriod=30 Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.611226 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.216:3000/\": read tcp 10.217.0.2:38770->10.217.0.216:3000: read: connection reset by peer" Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.930530 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.960622 4706 generic.go:334] "Generic (PLEG): container finished" podID="b6802220-f113-4842-b5d1-d7506f332423" containerID="49edd1079225936be761ace2702adc57714750e50e791a91da26884f0f914e17" exitCode=0 Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.960656 4706 generic.go:334] "Generic (PLEG): container finished" podID="b6802220-f113-4842-b5d1-d7506f332423" containerID="1d1ae3a8cdabebc3097b78f8d921797c22ef2a0442ee9adcfca703de4e2b2d73" exitCode=2 Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.960726 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerDied","Data":"49edd1079225936be761ace2702adc57714750e50e791a91da26884f0f914e17"} Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.960793 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerDied","Data":"1d1ae3a8cdabebc3097b78f8d921797c22ef2a0442ee9adcfca703de4e2b2d73"} Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.964862 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" event={"ID":"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa","Type":"ContainerStarted","Data":"8f4b403bc6c974c5ff8255c185cc2a11665f2efee113e234d62ef32b8d0a3a2e"} Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.964987 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-log" containerID="cri-o://36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267" gracePeriod=30 Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.965021 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.965093 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-api" containerID="cri-o://0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5" gracePeriod=30 Dec 08 19:45:49 crc kubenswrapper[4706]: I1208 19:45:49.991458 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" podStartSLOduration=2.991439125 podStartE2EDuration="2.991439125s" podCreationTimestamp="2025-12-08 19:45:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:49.989846949 +0000 UTC m=+1432.632047952" watchObservedRunningTime="2025-12-08 19:45:49.991439125 +0000 UTC m=+1432.633640128" Dec 08 19:45:50 crc kubenswrapper[4706]: I1208 19:45:50.979965 4706 generic.go:334] "Generic (PLEG): container finished" podID="928552a4-1547-43ac-afe5-645619776c5e" containerID="36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267" exitCode=143 Dec 08 19:45:50 crc kubenswrapper[4706]: I1208 19:45:50.980044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"928552a4-1547-43ac-afe5-645619776c5e","Type":"ContainerDied","Data":"36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267"} Dec 08 19:45:50 crc kubenswrapper[4706]: I1208 19:45:50.983728 4706 generic.go:334] "Generic (PLEG): container finished" podID="b6802220-f113-4842-b5d1-d7506f332423" containerID="950a12558264e5ceb64430e286b92d379cdd7a7add955288b0131f1dc80d2c55" exitCode=0 Dec 08 19:45:50 crc kubenswrapper[4706]: I1208 19:45:50.984877 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerDied","Data":"950a12558264e5ceb64430e286b92d379cdd7a7add955288b0131f1dc80d2c55"} Dec 08 19:45:51 crc kubenswrapper[4706]: I1208 19:45:51.270175 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 08 19:45:51 crc kubenswrapper[4706]: I1208 19:45:51.270604 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 08 19:45:51 crc kubenswrapper[4706]: I1208 19:45:51.300752 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:51 crc kubenswrapper[4706]: I1208 19:45:51.324407 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.016197 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.218437 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.216:3000/\": dial tcp 10.217.0.216:3000: connect: connection refused" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.272379 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-lp5nt"] Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.274565 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.278658 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.284362 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.285709 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lp5nt"] Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.300087 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.300621 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.455017 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-scripts\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.455321 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-config-data\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.455509 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.455766 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff6md\" (UniqueName: \"kubernetes.io/projected/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-kube-api-access-ff6md\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.559409 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-scripts\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.559663 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-config-data\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.559790 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.560060 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff6md\" (UniqueName: \"kubernetes.io/projected/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-kube-api-access-ff6md\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.568097 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.568128 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-scripts\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.573135 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-config-data\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.591948 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff6md\" (UniqueName: \"kubernetes.io/projected/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-kube-api-access-ff6md\") pod \"nova-cell1-cell-mapping-lp5nt\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:52 crc kubenswrapper[4706]: I1208 19:45:52.618176 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.216110 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lp5nt"] Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.704277 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.905614 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-combined-ca-bundle\") pod \"928552a4-1547-43ac-afe5-645619776c5e\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.906137 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-config-data\") pod \"928552a4-1547-43ac-afe5-645619776c5e\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.906247 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnphb\" (UniqueName: \"kubernetes.io/projected/928552a4-1547-43ac-afe5-645619776c5e-kube-api-access-fnphb\") pod \"928552a4-1547-43ac-afe5-645619776c5e\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.906370 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/928552a4-1547-43ac-afe5-645619776c5e-logs\") pod \"928552a4-1547-43ac-afe5-645619776c5e\" (UID: \"928552a4-1547-43ac-afe5-645619776c5e\") " Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.907963 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/928552a4-1547-43ac-afe5-645619776c5e-logs" (OuterVolumeSpecName: "logs") pod "928552a4-1547-43ac-afe5-645619776c5e" (UID: "928552a4-1547-43ac-afe5-645619776c5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.940592 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/928552a4-1547-43ac-afe5-645619776c5e-kube-api-access-fnphb" (OuterVolumeSpecName: "kube-api-access-fnphb") pod "928552a4-1547-43ac-afe5-645619776c5e" (UID: "928552a4-1547-43ac-afe5-645619776c5e"). InnerVolumeSpecName "kube-api-access-fnphb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.977458 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "928552a4-1547-43ac-afe5-645619776c5e" (UID: "928552a4-1547-43ac-afe5-645619776c5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:53 crc kubenswrapper[4706]: I1208 19:45:53.996012 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-config-data" (OuterVolumeSpecName: "config-data") pod "928552a4-1547-43ac-afe5-645619776c5e" (UID: "928552a4-1547-43ac-afe5-645619776c5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.009668 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/928552a4-1547-43ac-afe5-645619776c5e-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.009715 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.009733 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928552a4-1547-43ac-afe5-645619776c5e-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.009747 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnphb\" (UniqueName: \"kubernetes.io/projected/928552a4-1547-43ac-afe5-645619776c5e-kube-api-access-fnphb\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.034001 4706 generic.go:334] "Generic (PLEG): container finished" podID="b6802220-f113-4842-b5d1-d7506f332423" containerID="2f69d06b4582a22923544e57a10bef799b8c817f7847d022667b078daf68f25d" exitCode=0 Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.034098 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerDied","Data":"2f69d06b4582a22923544e57a10bef799b8c817f7847d022667b078daf68f25d"} Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.036165 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lp5nt" event={"ID":"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265","Type":"ContainerStarted","Data":"9274343acb4c2eaaaf30b25d2d2ed7b107a3f84c9522cecba5d8e1e10b88fab5"} Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.036201 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lp5nt" event={"ID":"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265","Type":"ContainerStarted","Data":"73ac92c7d6e9aa32cff37b83b2157b7b441f0ab1cd73528c4849b51a3de4d2bb"} Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.063551 4706 generic.go:334] "Generic (PLEG): container finished" podID="928552a4-1547-43ac-afe5-645619776c5e" containerID="0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5" exitCode=0 Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.063630 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"928552a4-1547-43ac-afe5-645619776c5e","Type":"ContainerDied","Data":"0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5"} Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.063663 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"928552a4-1547-43ac-afe5-645619776c5e","Type":"ContainerDied","Data":"4f799c7e4ddde50714733fd825c8d5bf65d6128301908570036f07abfb252c26"} Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.063707 4706 scope.go:117] "RemoveContainer" containerID="0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.063919 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.086959 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-lp5nt" podStartSLOduration=2.086923564 podStartE2EDuration="2.086923564s" podCreationTimestamp="2025-12-08 19:45:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:54.074601989 +0000 UTC m=+1436.716802992" watchObservedRunningTime="2025-12-08 19:45:54.086923564 +0000 UTC m=+1436.729124567" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.145173 4706 scope.go:117] "RemoveContainer" containerID="36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.157595 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.182607 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.194549 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:54 crc kubenswrapper[4706]: E1208 19:45:54.195434 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-api" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.195460 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-api" Dec 08 19:45:54 crc kubenswrapper[4706]: E1208 19:45:54.195502 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-log" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.195512 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-log" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.195801 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-api" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.195827 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="928552a4-1547-43ac-afe5-645619776c5e" containerName="nova-api-log" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.197721 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.204777 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.205171 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.205580 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.210818 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.219113 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bk7k\" (UniqueName: \"kubernetes.io/projected/8782e226-3bcf-47fd-927f-fa112bb686d0-kube-api-access-5bk7k\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.219342 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.219437 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-config-data\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.219465 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.219642 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.219939 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8782e226-3bcf-47fd-927f-fa112bb686d0-logs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.226530 4706 scope.go:117] "RemoveContainer" containerID="0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5" Dec 08 19:45:54 crc kubenswrapper[4706]: E1208 19:45:54.231088 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5\": container with ID starting with 0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5 not found: ID does not exist" containerID="0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.231531 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5"} err="failed to get container status \"0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5\": rpc error: code = NotFound desc = could not find container \"0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5\": container with ID starting with 0f712719f93bd9eea38de7ec6679ce167ca76a9e768100dbc1714509dde40ca5 not found: ID does not exist" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.231573 4706 scope.go:117] "RemoveContainer" containerID="36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267" Dec 08 19:45:54 crc kubenswrapper[4706]: E1208 19:45:54.232288 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267\": container with ID starting with 36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267 not found: ID does not exist" containerID="36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.232339 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267"} err="failed to get container status \"36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267\": rpc error: code = NotFound desc = could not find container \"36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267\": container with ID starting with 36f0452cb6e468b0102f8a2666adc02d5ca48c725c624586d54392d9db19d267 not found: ID does not exist" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.323281 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bk7k\" (UniqueName: \"kubernetes.io/projected/8782e226-3bcf-47fd-927f-fa112bb686d0-kube-api-access-5bk7k\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.323449 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.323517 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-config-data\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.323545 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.323638 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.323796 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8782e226-3bcf-47fd-927f-fa112bb686d0-logs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.324387 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8782e226-3bcf-47fd-927f-fa112bb686d0-logs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.340853 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.340916 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.340954 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.341239 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-config-data\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.351469 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bk7k\" (UniqueName: \"kubernetes.io/projected/8782e226-3bcf-47fd-927f-fa112bb686d0-kube-api-access-5bk7k\") pod \"nova-api-0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.470544 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.527954 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-config-data\") pod \"b6802220-f113-4842-b5d1-d7506f332423\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.528016 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwf8k\" (UniqueName: \"kubernetes.io/projected/b6802220-f113-4842-b5d1-d7506f332423-kube-api-access-rwf8k\") pod \"b6802220-f113-4842-b5d1-d7506f332423\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.528108 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-sg-core-conf-yaml\") pod \"b6802220-f113-4842-b5d1-d7506f332423\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.528152 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-scripts\") pod \"b6802220-f113-4842-b5d1-d7506f332423\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.528190 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-log-httpd\") pod \"b6802220-f113-4842-b5d1-d7506f332423\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.528284 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-ceilometer-tls-certs\") pod \"b6802220-f113-4842-b5d1-d7506f332423\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.528405 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-run-httpd\") pod \"b6802220-f113-4842-b5d1-d7506f332423\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.528588 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-combined-ca-bundle\") pod \"b6802220-f113-4842-b5d1-d7506f332423\" (UID: \"b6802220-f113-4842-b5d1-d7506f332423\") " Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.529427 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b6802220-f113-4842-b5d1-d7506f332423" (UID: "b6802220-f113-4842-b5d1-d7506f332423"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.534114 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-scripts" (OuterVolumeSpecName: "scripts") pod "b6802220-f113-4842-b5d1-d7506f332423" (UID: "b6802220-f113-4842-b5d1-d7506f332423"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.534434 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b6802220-f113-4842-b5d1-d7506f332423" (UID: "b6802220-f113-4842-b5d1-d7506f332423"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.536356 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6802220-f113-4842-b5d1-d7506f332423-kube-api-access-rwf8k" (OuterVolumeSpecName: "kube-api-access-rwf8k") pod "b6802220-f113-4842-b5d1-d7506f332423" (UID: "b6802220-f113-4842-b5d1-d7506f332423"). InnerVolumeSpecName "kube-api-access-rwf8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.538880 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.565821 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b6802220-f113-4842-b5d1-d7506f332423" (UID: "b6802220-f113-4842-b5d1-d7506f332423"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.602867 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b6802220-f113-4842-b5d1-d7506f332423" (UID: "b6802220-f113-4842-b5d1-d7506f332423"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.636561 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwf8k\" (UniqueName: \"kubernetes.io/projected/b6802220-f113-4842-b5d1-d7506f332423-kube-api-access-rwf8k\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.636609 4706 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.636623 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.636635 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.636649 4706 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.636662 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6802220-f113-4842-b5d1-d7506f332423-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.642738 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6802220-f113-4842-b5d1-d7506f332423" (UID: "b6802220-f113-4842-b5d1-d7506f332423"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.667504 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-config-data" (OuterVolumeSpecName: "config-data") pod "b6802220-f113-4842-b5d1-d7506f332423" (UID: "b6802220-f113-4842-b5d1-d7506f332423"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.750791 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:54 crc kubenswrapper[4706]: I1208 19:45:54.750859 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6802220-f113-4842-b5d1-d7506f332423-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.087961 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.088164 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6802220-f113-4842-b5d1-d7506f332423","Type":"ContainerDied","Data":"0294837cae3888419b2499a1d081632c0c482c7cc48ce4d6b273f84ae7e0f5da"} Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.088601 4706 scope.go:117] "RemoveContainer" containerID="49edd1079225936be761ace2702adc57714750e50e791a91da26884f0f914e17" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.142542 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.143931 4706 scope.go:117] "RemoveContainer" containerID="1d1ae3a8cdabebc3097b78f8d921797c22ef2a0442ee9adcfca703de4e2b2d73" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.167486 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.184543 4706 scope.go:117] "RemoveContainer" containerID="2f69d06b4582a22923544e57a10bef799b8c817f7847d022667b078daf68f25d" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.188938 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:55 crc kubenswrapper[4706]: E1208 19:45:55.189820 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="proxy-httpd" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.189852 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="proxy-httpd" Dec 08 19:45:55 crc kubenswrapper[4706]: E1208 19:45:55.189868 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="ceilometer-notification-agent" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.189877 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="ceilometer-notification-agent" Dec 08 19:45:55 crc kubenswrapper[4706]: E1208 19:45:55.189921 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="sg-core" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.189929 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="sg-core" Dec 08 19:45:55 crc kubenswrapper[4706]: E1208 19:45:55.189942 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="ceilometer-central-agent" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.189951 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="ceilometer-central-agent" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.190282 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="ceilometer-central-agent" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.190319 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="sg-core" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.190337 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="ceilometer-notification-agent" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.190360 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6802220-f113-4842-b5d1-d7506f332423" containerName="proxy-httpd" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.193644 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.213278 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.215629 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.215629 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.215915 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.239832 4706 scope.go:117] "RemoveContainer" containerID="950a12558264e5ceb64430e286b92d379cdd7a7add955288b0131f1dc80d2c55" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.253000 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.370965 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.371042 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-scripts\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.371095 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-log-httpd\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.371134 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-run-httpd\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.371240 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p9gt\" (UniqueName: \"kubernetes.io/projected/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-kube-api-access-9p9gt\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.371411 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.371458 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-config-data\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.371521 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473359 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-log-httpd\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473447 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-run-httpd\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473509 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p9gt\" (UniqueName: \"kubernetes.io/projected/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-kube-api-access-9p9gt\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473629 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473690 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-config-data\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473764 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473838 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-scripts\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.473944 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-log-httpd\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.475487 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-run-httpd\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.482158 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-scripts\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.482619 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.482788 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.483392 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-config-data\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.487012 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.498420 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p9gt\" (UniqueName: \"kubernetes.io/projected/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-kube-api-access-9p9gt\") pod \"ceilometer-0\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.599449 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.624200 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="928552a4-1547-43ac-afe5-645619776c5e" path="/var/lib/kubelet/pods/928552a4-1547-43ac-afe5-645619776c5e/volumes" Dec 08 19:45:55 crc kubenswrapper[4706]: I1208 19:45:55.624990 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6802220-f113-4842-b5d1-d7506f332423" path="/var/lib/kubelet/pods/b6802220-f113-4842-b5d1-d7506f332423/volumes" Dec 08 19:45:56 crc kubenswrapper[4706]: I1208 19:45:56.090070 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:45:56 crc kubenswrapper[4706]: W1208 19:45:56.100773 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf8a8bb4_8e88_42ab_9d29_ffd0a21732b3.slice/crio-ff2daf3d218a3cfcd10424467fe53ef814f7ec7be2c6cdc3339c53f334cdf55e WatchSource:0}: Error finding container ff2daf3d218a3cfcd10424467fe53ef814f7ec7be2c6cdc3339c53f334cdf55e: Status 404 returned error can't find the container with id ff2daf3d218a3cfcd10424467fe53ef814f7ec7be2c6cdc3339c53f334cdf55e Dec 08 19:45:56 crc kubenswrapper[4706]: I1208 19:45:56.121298 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8782e226-3bcf-47fd-927f-fa112bb686d0","Type":"ContainerStarted","Data":"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0"} Dec 08 19:45:56 crc kubenswrapper[4706]: I1208 19:45:56.121369 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8782e226-3bcf-47fd-927f-fa112bb686d0","Type":"ContainerStarted","Data":"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee"} Dec 08 19:45:56 crc kubenswrapper[4706]: I1208 19:45:56.121382 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8782e226-3bcf-47fd-927f-fa112bb686d0","Type":"ContainerStarted","Data":"a50d2e430c521b99f3ef5335372f0e06572f29754f44526e33afa74dcde2131e"} Dec 08 19:45:56 crc kubenswrapper[4706]: I1208 19:45:56.162143 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.162103105 podStartE2EDuration="2.162103105s" podCreationTimestamp="2025-12-08 19:45:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:45:56.145777624 +0000 UTC m=+1438.787978627" watchObservedRunningTime="2025-12-08 19:45:56.162103105 +0000 UTC m=+1438.804304108" Dec 08 19:45:57 crc kubenswrapper[4706]: I1208 19:45:57.137234 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerStarted","Data":"a9b012f8e4bd28755c8343dd753f4ff3e89351c2c6f7f6f8c2f90a34226aebdd"} Dec 08 19:45:57 crc kubenswrapper[4706]: I1208 19:45:57.139185 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerStarted","Data":"ff2daf3d218a3cfcd10424467fe53ef814f7ec7be2c6cdc3339c53f334cdf55e"} Dec 08 19:45:57 crc kubenswrapper[4706]: I1208 19:45:57.468492 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:45:57 crc kubenswrapper[4706]: I1208 19:45:57.573715 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-n79s4"] Dec 08 19:45:57 crc kubenswrapper[4706]: I1208 19:45:57.574369 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78cd565959-n79s4" podUID="deed8943-263a-449c-b219-72a8c0a838dd" containerName="dnsmasq-dns" containerID="cri-o://97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c" gracePeriod=10 Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.153772 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerStarted","Data":"5ecbee74c514a01e6b24dd30a90a858b4eda5fe2013ac12f82d328d7764e6ab0"} Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.157065 4706 generic.go:334] "Generic (PLEG): container finished" podID="deed8943-263a-449c-b219-72a8c0a838dd" containerID="97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c" exitCode=0 Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.157122 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-n79s4" event={"ID":"deed8943-263a-449c-b219-72a8c0a838dd","Type":"ContainerDied","Data":"97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c"} Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.346641 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.396168 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-swift-storage-0\") pod \"deed8943-263a-449c-b219-72a8c0a838dd\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.396445 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpv25\" (UniqueName: \"kubernetes.io/projected/deed8943-263a-449c-b219-72a8c0a838dd-kube-api-access-qpv25\") pod \"deed8943-263a-449c-b219-72a8c0a838dd\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.396629 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-svc\") pod \"deed8943-263a-449c-b219-72a8c0a838dd\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.396695 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-nb\") pod \"deed8943-263a-449c-b219-72a8c0a838dd\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.396745 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-sb\") pod \"deed8943-263a-449c-b219-72a8c0a838dd\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.397601 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-config\") pod \"deed8943-263a-449c-b219-72a8c0a838dd\" (UID: \"deed8943-263a-449c-b219-72a8c0a838dd\") " Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.417109 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deed8943-263a-449c-b219-72a8c0a838dd-kube-api-access-qpv25" (OuterVolumeSpecName: "kube-api-access-qpv25") pod "deed8943-263a-449c-b219-72a8c0a838dd" (UID: "deed8943-263a-449c-b219-72a8c0a838dd"). InnerVolumeSpecName "kube-api-access-qpv25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.528168 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpv25\" (UniqueName: \"kubernetes.io/projected/deed8943-263a-449c-b219-72a8c0a838dd-kube-api-access-qpv25\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.569520 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-config" (OuterVolumeSpecName: "config") pod "deed8943-263a-449c-b219-72a8c0a838dd" (UID: "deed8943-263a-449c-b219-72a8c0a838dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.584965 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "deed8943-263a-449c-b219-72a8c0a838dd" (UID: "deed8943-263a-449c-b219-72a8c0a838dd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.606718 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "deed8943-263a-449c-b219-72a8c0a838dd" (UID: "deed8943-263a-449c-b219-72a8c0a838dd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.615438 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "deed8943-263a-449c-b219-72a8c0a838dd" (UID: "deed8943-263a-449c-b219-72a8c0a838dd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.630411 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.630516 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.630533 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.630547 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.670175 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "deed8943-263a-449c-b219-72a8c0a838dd" (UID: "deed8943-263a-449c-b219-72a8c0a838dd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:45:58 crc kubenswrapper[4706]: I1208 19:45:58.733084 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deed8943-263a-449c-b219-72a8c0a838dd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:45:59 crc kubenswrapper[4706]: I1208 19:45:59.171028 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-n79s4" event={"ID":"deed8943-263a-449c-b219-72a8c0a838dd","Type":"ContainerDied","Data":"e001fc1c34a7726126caf28f85d86de0d5fce222e2e91eaa895a270dfcbfc81a"} Dec 08 19:45:59 crc kubenswrapper[4706]: I1208 19:45:59.171099 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-n79s4" Dec 08 19:45:59 crc kubenswrapper[4706]: I1208 19:45:59.171644 4706 scope.go:117] "RemoveContainer" containerID="97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c" Dec 08 19:45:59 crc kubenswrapper[4706]: I1208 19:45:59.214927 4706 scope.go:117] "RemoveContainer" containerID="4bd49a357b0e47975ef354cc243dc5989ab884ec4a5ae8464420afebff883e61" Dec 08 19:45:59 crc kubenswrapper[4706]: I1208 19:45:59.216065 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-n79s4"] Dec 08 19:45:59 crc kubenswrapper[4706]: I1208 19:45:59.231693 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-n79s4"] Dec 08 19:45:59 crc kubenswrapper[4706]: I1208 19:45:59.621498 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deed8943-263a-449c-b219-72a8c0a838dd" path="/var/lib/kubelet/pods/deed8943-263a-449c-b219-72a8c0a838dd/volumes" Dec 08 19:46:00 crc kubenswrapper[4706]: I1208 19:46:00.185747 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerStarted","Data":"1a66811b1cd05cf27bf4fe69f0869120c8952ee430cb1367f810ee022ac786b2"} Dec 08 19:46:01 crc kubenswrapper[4706]: I1208 19:46:01.202170 4706 generic.go:334] "Generic (PLEG): container finished" podID="b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" containerID="9274343acb4c2eaaaf30b25d2d2ed7b107a3f84c9522cecba5d8e1e10b88fab5" exitCode=0 Dec 08 19:46:01 crc kubenswrapper[4706]: I1208 19:46:01.202251 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lp5nt" event={"ID":"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265","Type":"ContainerDied","Data":"9274343acb4c2eaaaf30b25d2d2ed7b107a3f84c9522cecba5d8e1e10b88fab5"} Dec 08 19:46:01 crc kubenswrapper[4706]: I1208 19:46:01.234692 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerStarted","Data":"6f01a8aeeb93a9681f3f65c2b065510158b3d67a4a1f6d3ac493ee976f665dbf"} Dec 08 19:46:01 crc kubenswrapper[4706]: I1208 19:46:01.236227 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 19:46:01 crc kubenswrapper[4706]: I1208 19:46:01.269176 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9400055950000001 podStartE2EDuration="6.269153185s" podCreationTimestamp="2025-12-08 19:45:55 +0000 UTC" firstStartedPulling="2025-12-08 19:45:56.106053598 +0000 UTC m=+1438.748254601" lastFinishedPulling="2025-12-08 19:46:00.435201168 +0000 UTC m=+1443.077402191" observedRunningTime="2025-12-08 19:46:01.256780768 +0000 UTC m=+1443.898981791" watchObservedRunningTime="2025-12-08 19:46:01.269153185 +0000 UTC m=+1443.911354188" Dec 08 19:46:01 crc kubenswrapper[4706]: I1208 19:46:01.276447 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 08 19:46:01 crc kubenswrapper[4706]: I1208 19:46:01.282899 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 08 19:46:01 crc kubenswrapper[4706]: I1208 19:46:01.285947 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 08 19:46:02 crc kubenswrapper[4706]: I1208 19:46:02.258715 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 08 19:46:02 crc kubenswrapper[4706]: I1208 19:46:02.846057 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:46:02 crc kubenswrapper[4706]: I1208 19:46:02.984711 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff6md\" (UniqueName: \"kubernetes.io/projected/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-kube-api-access-ff6md\") pod \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " Dec 08 19:46:02 crc kubenswrapper[4706]: I1208 19:46:02.984776 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-scripts\") pod \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " Dec 08 19:46:02 crc kubenswrapper[4706]: I1208 19:46:02.984831 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-config-data\") pod \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " Dec 08 19:46:02 crc kubenswrapper[4706]: I1208 19:46:02.986087 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-combined-ca-bundle\") pod \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\" (UID: \"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265\") " Dec 08 19:46:02 crc kubenswrapper[4706]: I1208 19:46:02.993715 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-kube-api-access-ff6md" (OuterVolumeSpecName: "kube-api-access-ff6md") pod "b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" (UID: "b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265"). InnerVolumeSpecName "kube-api-access-ff6md". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:46:02 crc kubenswrapper[4706]: I1208 19:46:02.994752 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-scripts" (OuterVolumeSpecName: "scripts") pod "b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" (UID: "b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.021448 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" (UID: "b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.033424 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-config-data" (OuterVolumeSpecName: "config-data") pod "b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" (UID: "b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.088709 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.088758 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff6md\" (UniqueName: \"kubernetes.io/projected/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-kube-api-access-ff6md\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.088772 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.088909 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.265102 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lp5nt" event={"ID":"b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265","Type":"ContainerDied","Data":"73ac92c7d6e9aa32cff37b83b2157b7b441f0ab1cd73528c4849b51a3de4d2bb"} Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.265168 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73ac92c7d6e9aa32cff37b83b2157b7b441f0ab1cd73528c4849b51a3de4d2bb" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.265575 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lp5nt" Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.457589 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.458064 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerName="nova-api-log" containerID="cri-o://58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee" gracePeriod=30 Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.458475 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerName="nova-api-api" containerID="cri-o://933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0" gracePeriod=30 Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.489307 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.489580 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="713c3abb-338f-4a61-a52d-c3447a7dceec" containerName="nova-scheduler-scheduler" containerID="cri-o://140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc" gracePeriod=30 Dec 08 19:46:03 crc kubenswrapper[4706]: I1208 19:46:03.504787 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:46:03 crc kubenswrapper[4706]: E1208 19:46:03.902657 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 19:46:03 crc kubenswrapper[4706]: E1208 19:46:03.907479 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 19:46:03 crc kubenswrapper[4706]: E1208 19:46:03.928444 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 19:46:03 crc kubenswrapper[4706]: E1208 19:46:03.928583 4706 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="713c3abb-338f-4a61-a52d-c3447a7dceec" containerName="nova-scheduler-scheduler" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.218013 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.289164 4706 generic.go:334] "Generic (PLEG): container finished" podID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerID="933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0" exitCode=0 Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.289200 4706 generic.go:334] "Generic (PLEG): container finished" podID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerID="58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee" exitCode=143 Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.289494 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8782e226-3bcf-47fd-927f-fa112bb686d0","Type":"ContainerDied","Data":"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0"} Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.289549 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8782e226-3bcf-47fd-927f-fa112bb686d0","Type":"ContainerDied","Data":"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee"} Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.289563 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8782e226-3bcf-47fd-927f-fa112bb686d0","Type":"ContainerDied","Data":"a50d2e430c521b99f3ef5335372f0e06572f29754f44526e33afa74dcde2131e"} Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.289584 4706 scope.go:117] "RemoveContainer" containerID="933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.289647 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.323416 4706 scope.go:117] "RemoveContainer" containerID="58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.324026 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-combined-ca-bundle\") pod \"8782e226-3bcf-47fd-927f-fa112bb686d0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.324087 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-internal-tls-certs\") pod \"8782e226-3bcf-47fd-927f-fa112bb686d0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.324158 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-config-data\") pod \"8782e226-3bcf-47fd-927f-fa112bb686d0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.324288 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bk7k\" (UniqueName: \"kubernetes.io/projected/8782e226-3bcf-47fd-927f-fa112bb686d0-kube-api-access-5bk7k\") pod \"8782e226-3bcf-47fd-927f-fa112bb686d0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.324400 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-public-tls-certs\") pod \"8782e226-3bcf-47fd-927f-fa112bb686d0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.324475 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8782e226-3bcf-47fd-927f-fa112bb686d0-logs\") pod \"8782e226-3bcf-47fd-927f-fa112bb686d0\" (UID: \"8782e226-3bcf-47fd-927f-fa112bb686d0\") " Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.325530 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8782e226-3bcf-47fd-927f-fa112bb686d0-logs" (OuterVolumeSpecName: "logs") pod "8782e226-3bcf-47fd-927f-fa112bb686d0" (UID: "8782e226-3bcf-47fd-927f-fa112bb686d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.350939 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8782e226-3bcf-47fd-927f-fa112bb686d0-kube-api-access-5bk7k" (OuterVolumeSpecName: "kube-api-access-5bk7k") pod "8782e226-3bcf-47fd-927f-fa112bb686d0" (UID: "8782e226-3bcf-47fd-927f-fa112bb686d0"). InnerVolumeSpecName "kube-api-access-5bk7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.374532 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-config-data" (OuterVolumeSpecName: "config-data") pod "8782e226-3bcf-47fd-927f-fa112bb686d0" (UID: "8782e226-3bcf-47fd-927f-fa112bb686d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.379520 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8782e226-3bcf-47fd-927f-fa112bb686d0" (UID: "8782e226-3bcf-47fd-927f-fa112bb686d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.410788 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8782e226-3bcf-47fd-927f-fa112bb686d0" (UID: "8782e226-3bcf-47fd-927f-fa112bb686d0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.422578 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8782e226-3bcf-47fd-927f-fa112bb686d0" (UID: "8782e226-3bcf-47fd-927f-fa112bb686d0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.436395 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bk7k\" (UniqueName: \"kubernetes.io/projected/8782e226-3bcf-47fd-927f-fa112bb686d0-kube-api-access-5bk7k\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.436445 4706 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.436460 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8782e226-3bcf-47fd-927f-fa112bb686d0-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.436471 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.436480 4706 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.436490 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8782e226-3bcf-47fd-927f-fa112bb686d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.501220 4706 scope.go:117] "RemoveContainer" containerID="933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0" Dec 08 19:46:04 crc kubenswrapper[4706]: E1208 19:46:04.502448 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0\": container with ID starting with 933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0 not found: ID does not exist" containerID="933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.502493 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0"} err="failed to get container status \"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0\": rpc error: code = NotFound desc = could not find container \"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0\": container with ID starting with 933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0 not found: ID does not exist" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.502525 4706 scope.go:117] "RemoveContainer" containerID="58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee" Dec 08 19:46:04 crc kubenswrapper[4706]: E1208 19:46:04.502987 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee\": container with ID starting with 58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee not found: ID does not exist" containerID="58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.503023 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee"} err="failed to get container status \"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee\": rpc error: code = NotFound desc = could not find container \"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee\": container with ID starting with 58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee not found: ID does not exist" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.503045 4706 scope.go:117] "RemoveContainer" containerID="933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.503527 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0"} err="failed to get container status \"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0\": rpc error: code = NotFound desc = could not find container \"933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0\": container with ID starting with 933086e2265ce2e10223be7fd2d447b5df98d92d940a2015e25362695692c7a0 not found: ID does not exist" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.503639 4706 scope.go:117] "RemoveContainer" containerID="58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.504189 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee"} err="failed to get container status \"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee\": rpc error: code = NotFound desc = could not find container \"58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee\": container with ID starting with 58cfc0ebe2c2d47b324b539eb43651b13cc803b86fbf0f50699860223067bdee not found: ID does not exist" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.671226 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.699011 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.713990 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 08 19:46:04 crc kubenswrapper[4706]: E1208 19:46:04.714720 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerName="nova-api-api" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.714749 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerName="nova-api-api" Dec 08 19:46:04 crc kubenswrapper[4706]: E1208 19:46:04.714786 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deed8943-263a-449c-b219-72a8c0a838dd" containerName="init" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.714795 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="deed8943-263a-449c-b219-72a8c0a838dd" containerName="init" Dec 08 19:46:04 crc kubenswrapper[4706]: E1208 19:46:04.714811 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" containerName="nova-manage" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.714817 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" containerName="nova-manage" Dec 08 19:46:04 crc kubenswrapper[4706]: E1208 19:46:04.714834 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerName="nova-api-log" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.714844 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerName="nova-api-log" Dec 08 19:46:04 crc kubenswrapper[4706]: E1208 19:46:04.714867 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deed8943-263a-449c-b219-72a8c0a838dd" containerName="dnsmasq-dns" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.714876 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="deed8943-263a-449c-b219-72a8c0a838dd" containerName="dnsmasq-dns" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.715118 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" containerName="nova-manage" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.715141 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="deed8943-263a-449c-b219-72a8c0a838dd" containerName="dnsmasq-dns" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.715162 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerName="nova-api-api" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.715171 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" containerName="nova-api-log" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.717009 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.725497 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.726883 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.727070 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.728359 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.849443 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.849517 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-config-data\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.849642 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.849685 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48240f0b-4bff-4e59-b151-7e2922ee95f0-logs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.849792 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-public-tls-certs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.849812 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfndd\" (UniqueName: \"kubernetes.io/projected/48240f0b-4bff-4e59-b151-7e2922ee95f0-kube-api-access-jfndd\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.954144 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-public-tls-certs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.954196 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfndd\" (UniqueName: \"kubernetes.io/projected/48240f0b-4bff-4e59-b151-7e2922ee95f0-kube-api-access-jfndd\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.954318 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.954365 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-config-data\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.954457 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.954479 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48240f0b-4bff-4e59-b151-7e2922ee95f0-logs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.955086 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48240f0b-4bff-4e59-b151-7e2922ee95f0-logs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.970080 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-public-tls-certs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.973721 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-config-data\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.973840 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.975168 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48240f0b-4bff-4e59-b151-7e2922ee95f0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:04 crc kubenswrapper[4706]: I1208 19:46:04.978867 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfndd\" (UniqueName: \"kubernetes.io/projected/48240f0b-4bff-4e59-b151-7e2922ee95f0-kube-api-access-jfndd\") pod \"nova-api-0\" (UID: \"48240f0b-4bff-4e59-b151-7e2922ee95f0\") " pod="openstack/nova-api-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.124167 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.225628 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.307466 4706 generic.go:334] "Generic (PLEG): container finished" podID="713c3abb-338f-4a61-a52d-c3447a7dceec" containerID="140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc" exitCode=0 Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.307608 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"713c3abb-338f-4a61-a52d-c3447a7dceec","Type":"ContainerDied","Data":"140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc"} Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.308857 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"713c3abb-338f-4a61-a52d-c3447a7dceec","Type":"ContainerDied","Data":"305bdb21bc95d315149718a962ea761d3710aaaac399187a2460bae2af35d554"} Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.308894 4706 scope.go:117] "RemoveContainer" containerID="140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.307702 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.309044 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-log" containerID="cri-o://a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180" gracePeriod=30 Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.309105 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-metadata" containerID="cri-o://20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0" gracePeriod=30 Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.367422 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-combined-ca-bundle\") pod \"713c3abb-338f-4a61-a52d-c3447a7dceec\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.367486 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-config-data\") pod \"713c3abb-338f-4a61-a52d-c3447a7dceec\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.367677 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8867d\" (UniqueName: \"kubernetes.io/projected/713c3abb-338f-4a61-a52d-c3447a7dceec-kube-api-access-8867d\") pod \"713c3abb-338f-4a61-a52d-c3447a7dceec\" (UID: \"713c3abb-338f-4a61-a52d-c3447a7dceec\") " Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.379508 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/713c3abb-338f-4a61-a52d-c3447a7dceec-kube-api-access-8867d" (OuterVolumeSpecName: "kube-api-access-8867d") pod "713c3abb-338f-4a61-a52d-c3447a7dceec" (UID: "713c3abb-338f-4a61-a52d-c3447a7dceec"). InnerVolumeSpecName "kube-api-access-8867d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.387649 4706 scope.go:117] "RemoveContainer" containerID="140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc" Dec 08 19:46:05 crc kubenswrapper[4706]: E1208 19:46:05.391529 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc\": container with ID starting with 140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc not found: ID does not exist" containerID="140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.391596 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc"} err="failed to get container status \"140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc\": rpc error: code = NotFound desc = could not find container \"140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc\": container with ID starting with 140c61bae63d22786b5bb11bd30412262ee6fa2562b424cbe3518bfc9252d9fc not found: ID does not exist" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.415650 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "713c3abb-338f-4a61-a52d-c3447a7dceec" (UID: "713c3abb-338f-4a61-a52d-c3447a7dceec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.432169 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-config-data" (OuterVolumeSpecName: "config-data") pod "713c3abb-338f-4a61-a52d-c3447a7dceec" (UID: "713c3abb-338f-4a61-a52d-c3447a7dceec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.470974 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.471018 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713c3abb-338f-4a61-a52d-c3447a7dceec-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.471029 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8867d\" (UniqueName: \"kubernetes.io/projected/713c3abb-338f-4a61-a52d-c3447a7dceec-kube-api-access-8867d\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.628085 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8782e226-3bcf-47fd-927f-fa112bb686d0" path="/var/lib/kubelet/pods/8782e226-3bcf-47fd-927f-fa112bb686d0/volumes" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.661978 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.681551 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.696708 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:46:05 crc kubenswrapper[4706]: E1208 19:46:05.697463 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="713c3abb-338f-4a61-a52d-c3447a7dceec" containerName="nova-scheduler-scheduler" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.697503 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="713c3abb-338f-4a61-a52d-c3447a7dceec" containerName="nova-scheduler-scheduler" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.697832 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="713c3abb-338f-4a61-a52d-c3447a7dceec" containerName="nova-scheduler-scheduler" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.699010 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.702290 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.725122 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.745739 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.778705 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ae432b-2caa-458f-9ac4-5172e21ef898-config-data\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.779116 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ae432b-2caa-458f-9ac4-5172e21ef898-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.779361 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p24vn\" (UniqueName: \"kubernetes.io/projected/a0ae432b-2caa-458f-9ac4-5172e21ef898-kube-api-access-p24vn\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.836051 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.836110 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.881518 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ae432b-2caa-458f-9ac4-5172e21ef898-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.882070 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p24vn\" (UniqueName: \"kubernetes.io/projected/a0ae432b-2caa-458f-9ac4-5172e21ef898-kube-api-access-p24vn\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.882229 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ae432b-2caa-458f-9ac4-5172e21ef898-config-data\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.904635 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ae432b-2caa-458f-9ac4-5172e21ef898-config-data\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.905897 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ae432b-2caa-458f-9ac4-5172e21ef898-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:05 crc kubenswrapper[4706]: I1208 19:46:05.910656 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p24vn\" (UniqueName: \"kubernetes.io/projected/a0ae432b-2caa-458f-9ac4-5172e21ef898-kube-api-access-p24vn\") pod \"nova-scheduler-0\" (UID: \"a0ae432b-2caa-458f-9ac4-5172e21ef898\") " pod="openstack/nova-scheduler-0" Dec 08 19:46:06 crc kubenswrapper[4706]: I1208 19:46:06.157574 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 19:46:06 crc kubenswrapper[4706]: I1208 19:46:06.340355 4706 generic.go:334] "Generic (PLEG): container finished" podID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerID="a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180" exitCode=143 Dec 08 19:46:06 crc kubenswrapper[4706]: I1208 19:46:06.340440 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a77727be-8bf5-4c35-827b-ce13e018eaee","Type":"ContainerDied","Data":"a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180"} Dec 08 19:46:06 crc kubenswrapper[4706]: I1208 19:46:06.352572 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48240f0b-4bff-4e59-b151-7e2922ee95f0","Type":"ContainerStarted","Data":"e889b7506a4af74478f70cb46f3d70eb4acbccd9796aa3989263abcf1f7dc3e8"} Dec 08 19:46:06 crc kubenswrapper[4706]: I1208 19:46:06.352646 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48240f0b-4bff-4e59-b151-7e2922ee95f0","Type":"ContainerStarted","Data":"4404490d418cf5756c70b8159e66c828743b7b2be280dc4c034189bbdec7114b"} Dec 08 19:46:06 crc kubenswrapper[4706]: I1208 19:46:06.677169 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 19:46:06 crc kubenswrapper[4706]: W1208 19:46:06.679968 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0ae432b_2caa_458f_9ac4_5172e21ef898.slice/crio-b8d006918876c743ec8f7bcf6b326e05dfdec6e7ff26364beed67c1db0cb5acd WatchSource:0}: Error finding container b8d006918876c743ec8f7bcf6b326e05dfdec6e7ff26364beed67c1db0cb5acd: Status 404 returned error can't find the container with id b8d006918876c743ec8f7bcf6b326e05dfdec6e7ff26364beed67c1db0cb5acd Dec 08 19:46:06 crc kubenswrapper[4706]: E1208 19:46:06.910591 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeed8943_263a_449c_b219_72a8c0a838dd.slice/crio-conmon-97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c.scope\": RecentStats: unable to find data in memory cache]" Dec 08 19:46:07 crc kubenswrapper[4706]: I1208 19:46:07.370928 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a0ae432b-2caa-458f-9ac4-5172e21ef898","Type":"ContainerStarted","Data":"b8d006918876c743ec8f7bcf6b326e05dfdec6e7ff26364beed67c1db0cb5acd"} Dec 08 19:46:07 crc kubenswrapper[4706]: I1208 19:46:07.374099 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"48240f0b-4bff-4e59-b151-7e2922ee95f0","Type":"ContainerStarted","Data":"57f614bbe89db265abeaf7749a69173d7f9e7aefdc06f6d9ae042fbed8f366b3"} Dec 08 19:46:07 crc kubenswrapper[4706]: I1208 19:46:07.399457 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.39942223 podStartE2EDuration="3.39942223s" podCreationTimestamp="2025-12-08 19:46:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:46:07.39458209 +0000 UTC m=+1450.036783093" watchObservedRunningTime="2025-12-08 19:46:07.39942223 +0000 UTC m=+1450.041623233" Dec 08 19:46:07 crc kubenswrapper[4706]: I1208 19:46:07.622128 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="713c3abb-338f-4a61-a52d-c3447a7dceec" path="/var/lib/kubelet/pods/713c3abb-338f-4a61-a52d-c3447a7dceec/volumes" Dec 08 19:46:08 crc kubenswrapper[4706]: I1208 19:46:08.412302 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a0ae432b-2caa-458f-9ac4-5172e21ef898","Type":"ContainerStarted","Data":"fffbeaa84fcc206e09d2f46173ccea1130e67c1309ba79c25f75269e8a8b16d4"} Dec 08 19:46:08 crc kubenswrapper[4706]: I1208 19:46:08.437113 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.437086122 podStartE2EDuration="3.437086122s" podCreationTimestamp="2025-12-08 19:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:46:08.430186903 +0000 UTC m=+1451.072387896" watchObservedRunningTime="2025-12-08 19:46:08.437086122 +0000 UTC m=+1451.079287125" Dec 08 19:46:08 crc kubenswrapper[4706]: I1208 19:46:08.486646 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": read tcp 10.217.0.2:58952->10.217.0.219:8775: read: connection reset by peer" Dec 08 19:46:08 crc kubenswrapper[4706]: I1208 19:46:08.486996 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": read tcp 10.217.0.2:58944->10.217.0.219:8775: read: connection reset by peer" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.312153 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.423150 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-config-data\") pod \"a77727be-8bf5-4c35-827b-ce13e018eaee\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.423346 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-nova-metadata-tls-certs\") pod \"a77727be-8bf5-4c35-827b-ce13e018eaee\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.423515 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-combined-ca-bundle\") pod \"a77727be-8bf5-4c35-827b-ce13e018eaee\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.423676 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmh7w\" (UniqueName: \"kubernetes.io/projected/a77727be-8bf5-4c35-827b-ce13e018eaee-kube-api-access-xmh7w\") pod \"a77727be-8bf5-4c35-827b-ce13e018eaee\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.423712 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a77727be-8bf5-4c35-827b-ce13e018eaee-logs\") pod \"a77727be-8bf5-4c35-827b-ce13e018eaee\" (UID: \"a77727be-8bf5-4c35-827b-ce13e018eaee\") " Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.425740 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a77727be-8bf5-4c35-827b-ce13e018eaee-logs" (OuterVolumeSpecName: "logs") pod "a77727be-8bf5-4c35-827b-ce13e018eaee" (UID: "a77727be-8bf5-4c35-827b-ce13e018eaee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.429710 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a77727be-8bf5-4c35-827b-ce13e018eaee-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.441561 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a77727be-8bf5-4c35-827b-ce13e018eaee-kube-api-access-xmh7w" (OuterVolumeSpecName: "kube-api-access-xmh7w") pod "a77727be-8bf5-4c35-827b-ce13e018eaee" (UID: "a77727be-8bf5-4c35-827b-ce13e018eaee"). InnerVolumeSpecName "kube-api-access-xmh7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.445211 4706 generic.go:334] "Generic (PLEG): container finished" podID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerID="20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0" exitCode=0 Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.446823 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.447195 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a77727be-8bf5-4c35-827b-ce13e018eaee","Type":"ContainerDied","Data":"20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0"} Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.447360 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a77727be-8bf5-4c35-827b-ce13e018eaee","Type":"ContainerDied","Data":"47e8fb021e3e9191fd0d8bc4c906964991b00d68c8582c3bbe4fcef4487abd37"} Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.447457 4706 scope.go:117] "RemoveContainer" containerID="20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.485866 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-config-data" (OuterVolumeSpecName: "config-data") pod "a77727be-8bf5-4c35-827b-ce13e018eaee" (UID: "a77727be-8bf5-4c35-827b-ce13e018eaee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.510607 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a77727be-8bf5-4c35-827b-ce13e018eaee" (UID: "a77727be-8bf5-4c35-827b-ce13e018eaee"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.532323 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a77727be-8bf5-4c35-827b-ce13e018eaee" (UID: "a77727be-8bf5-4c35-827b-ce13e018eaee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.532783 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.532899 4706 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.532973 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a77727be-8bf5-4c35-827b-ce13e018eaee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.533031 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmh7w\" (UniqueName: \"kubernetes.io/projected/a77727be-8bf5-4c35-827b-ce13e018eaee-kube-api-access-xmh7w\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.672562 4706 scope.go:117] "RemoveContainer" containerID="a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.712710 4706 scope.go:117] "RemoveContainer" containerID="20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0" Dec 08 19:46:09 crc kubenswrapper[4706]: E1208 19:46:09.714188 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0\": container with ID starting with 20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0 not found: ID does not exist" containerID="20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.714234 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0"} err="failed to get container status \"20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0\": rpc error: code = NotFound desc = could not find container \"20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0\": container with ID starting with 20cdaa95b9f3f6c983fa99c53fb9a228597557e4cb2fc08715556175dcedb0c0 not found: ID does not exist" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.714285 4706 scope.go:117] "RemoveContainer" containerID="a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180" Dec 08 19:46:09 crc kubenswrapper[4706]: E1208 19:46:09.716055 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180\": container with ID starting with a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180 not found: ID does not exist" containerID="a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.716080 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180"} err="failed to get container status \"a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180\": rpc error: code = NotFound desc = could not find container \"a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180\": container with ID starting with a182d9561edf2c8a8af74e7cc9c97b39dfa8373855f14feda52d603249169180 not found: ID does not exist" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.787499 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.813837 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.843299 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:46:09 crc kubenswrapper[4706]: E1208 19:46:09.844338 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-log" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.844439 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-log" Dec 08 19:46:09 crc kubenswrapper[4706]: E1208 19:46:09.844495 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-metadata" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.844544 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-metadata" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.844826 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-metadata" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.844907 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" containerName="nova-metadata-log" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.846416 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.855126 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.856119 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.856354 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.946241 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-config-data\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.946691 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.946870 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2xlz\" (UniqueName: \"kubernetes.io/projected/7eb161ee-d8ee-45e4-9fce-a0661293603a-kube-api-access-w2xlz\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.947006 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:09 crc kubenswrapper[4706]: I1208 19:46:09.947098 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eb161ee-d8ee-45e4-9fce-a0661293603a-logs\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.049146 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eb161ee-d8ee-45e4-9fce-a0661293603a-logs\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.049274 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-config-data\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.049396 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.049704 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2xlz\" (UniqueName: \"kubernetes.io/projected/7eb161ee-d8ee-45e4-9fce-a0661293603a-kube-api-access-w2xlz\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.049787 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.051347 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eb161ee-d8ee-45e4-9fce-a0661293603a-logs\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.055837 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.056835 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.057526 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb161ee-d8ee-45e4-9fce-a0661293603a-config-data\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.071115 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2xlz\" (UniqueName: \"kubernetes.io/projected/7eb161ee-d8ee-45e4-9fce-a0661293603a-kube-api-access-w2xlz\") pod \"nova-metadata-0\" (UID: \"7eb161ee-d8ee-45e4-9fce-a0661293603a\") " pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.176868 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 19:46:10 crc kubenswrapper[4706]: I1208 19:46:10.668982 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 19:46:10 crc kubenswrapper[4706]: W1208 19:46:10.669071 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eb161ee_d8ee_45e4_9fce_a0661293603a.slice/crio-c3ec16ace4e94d9cf520ae8542512c9554fd62146b00db515861669b967a5328 WatchSource:0}: Error finding container c3ec16ace4e94d9cf520ae8542512c9554fd62146b00db515861669b967a5328: Status 404 returned error can't find the container with id c3ec16ace4e94d9cf520ae8542512c9554fd62146b00db515861669b967a5328 Dec 08 19:46:11 crc kubenswrapper[4706]: I1208 19:46:11.158034 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 08 19:46:11 crc kubenswrapper[4706]: I1208 19:46:11.479809 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7eb161ee-d8ee-45e4-9fce-a0661293603a","Type":"ContainerStarted","Data":"e40fa41761d547844ec726e788f624e2cc829dd2151f3f714f84d6604a9a07ac"} Dec 08 19:46:11 crc kubenswrapper[4706]: I1208 19:46:11.480348 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7eb161ee-d8ee-45e4-9fce-a0661293603a","Type":"ContainerStarted","Data":"4878c202db91cc13d3d8fc2e0c1c2a3ccd71cccdded1da324537edce7e64dff7"} Dec 08 19:46:11 crc kubenswrapper[4706]: I1208 19:46:11.480363 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7eb161ee-d8ee-45e4-9fce-a0661293603a","Type":"ContainerStarted","Data":"c3ec16ace4e94d9cf520ae8542512c9554fd62146b00db515861669b967a5328"} Dec 08 19:46:11 crc kubenswrapper[4706]: I1208 19:46:11.504233 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.504197907 podStartE2EDuration="2.504197907s" podCreationTimestamp="2025-12-08 19:46:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:46:11.499656696 +0000 UTC m=+1454.141857699" watchObservedRunningTime="2025-12-08 19:46:11.504197907 +0000 UTC m=+1454.146398910" Dec 08 19:46:11 crc kubenswrapper[4706]: I1208 19:46:11.624029 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a77727be-8bf5-4c35-827b-ce13e018eaee" path="/var/lib/kubelet/pods/a77727be-8bf5-4c35-827b-ce13e018eaee/volumes" Dec 08 19:46:15 crc kubenswrapper[4706]: I1208 19:46:15.125223 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 19:46:15 crc kubenswrapper[4706]: I1208 19:46:15.126011 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 19:46:15 crc kubenswrapper[4706]: I1208 19:46:15.177336 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 19:46:15 crc kubenswrapper[4706]: I1208 19:46:15.177394 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 19:46:16 crc kubenswrapper[4706]: I1208 19:46:16.139527 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="48240f0b-4bff-4e59-b151-7e2922ee95f0" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 19:46:16 crc kubenswrapper[4706]: I1208 19:46:16.139582 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="48240f0b-4bff-4e59-b151-7e2922ee95f0" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 19:46:16 crc kubenswrapper[4706]: I1208 19:46:16.158339 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 08 19:46:16 crc kubenswrapper[4706]: I1208 19:46:16.348744 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 08 19:46:16 crc kubenswrapper[4706]: I1208 19:46:16.581247 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 08 19:46:17 crc kubenswrapper[4706]: E1208 19:46:17.255559 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeed8943_263a_449c_b219_72a8c0a838dd.slice/crio-conmon-97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c.scope\": RecentStats: unable to find data in memory cache]" Dec 08 19:46:20 crc kubenswrapper[4706]: I1208 19:46:20.177734 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 08 19:46:20 crc kubenswrapper[4706]: I1208 19:46:20.178657 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 08 19:46:21 crc kubenswrapper[4706]: I1208 19:46:21.190403 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7eb161ee-d8ee-45e4-9fce-a0661293603a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.227:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 19:46:21 crc kubenswrapper[4706]: I1208 19:46:21.190683 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7eb161ee-d8ee-45e4-9fce-a0661293603a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.227:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 19:46:25 crc kubenswrapper[4706]: I1208 19:46:25.133975 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 08 19:46:25 crc kubenswrapper[4706]: I1208 19:46:25.135590 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 08 19:46:25 crc kubenswrapper[4706]: I1208 19:46:25.139613 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 08 19:46:25 crc kubenswrapper[4706]: I1208 19:46:25.145860 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 08 19:46:25 crc kubenswrapper[4706]: I1208 19:46:25.621789 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 08 19:46:25 crc kubenswrapper[4706]: I1208 19:46:25.677280 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 08 19:46:25 crc kubenswrapper[4706]: I1208 19:46:25.686020 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 08 19:46:27 crc kubenswrapper[4706]: E1208 19:46:27.555842 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeed8943_263a_449c_b219_72a8c0a838dd.slice/crio-conmon-97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c.scope\": RecentStats: unable to find data in memory cache]" Dec 08 19:46:30 crc kubenswrapper[4706]: I1208 19:46:30.198772 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 08 19:46:30 crc kubenswrapper[4706]: I1208 19:46:30.199631 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 08 19:46:30 crc kubenswrapper[4706]: I1208 19:46:30.205814 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 08 19:46:30 crc kubenswrapper[4706]: I1208 19:46:30.206245 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 08 19:46:35 crc kubenswrapper[4706]: I1208 19:46:35.835558 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:46:35 crc kubenswrapper[4706]: I1208 19:46:35.836504 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:46:35 crc kubenswrapper[4706]: I1208 19:46:35.836571 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:46:35 crc kubenswrapper[4706]: I1208 19:46:35.837657 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a5681b5fe1d33f6e2441d03b7699cdb0e62dc0c63f422690f0cadd789557d42f"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:46:35 crc kubenswrapper[4706]: I1208 19:46:35.837723 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://a5681b5fe1d33f6e2441d03b7699cdb0e62dc0c63f422690f0cadd789557d42f" gracePeriod=600 Dec 08 19:46:36 crc kubenswrapper[4706]: I1208 19:46:36.812892 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="a5681b5fe1d33f6e2441d03b7699cdb0e62dc0c63f422690f0cadd789557d42f" exitCode=0 Dec 08 19:46:36 crc kubenswrapper[4706]: I1208 19:46:36.813685 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"a5681b5fe1d33f6e2441d03b7699cdb0e62dc0c63f422690f0cadd789557d42f"} Dec 08 19:46:36 crc kubenswrapper[4706]: I1208 19:46:36.813721 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494"} Dec 08 19:46:36 crc kubenswrapper[4706]: I1208 19:46:36.813746 4706 scope.go:117] "RemoveContainer" containerID="5cdb7c8ca56123f7a237303ad544e086627575cfd4a9196d909f15ed3eb4ca79" Dec 08 19:46:38 crc kubenswrapper[4706]: E1208 19:46:38.049046 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeed8943_263a_449c_b219_72a8c0a838dd.slice/crio-conmon-97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c.scope\": RecentStats: unable to find data in memory cache]" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.514964 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-59fls"] Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.528063 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-59fls"] Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.626588 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-wflqk"] Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.628296 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.636170 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.643423 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-wflqk"] Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.733861 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-combined-ca-bundle\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.734131 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-certs\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.734242 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lpmw\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-kube-api-access-7lpmw\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.734300 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-config-data\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.734377 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-scripts\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.838283 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-combined-ca-bundle\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.838404 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-certs\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.838462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lpmw\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-kube-api-access-7lpmw\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.838484 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-config-data\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.838520 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-scripts\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.848296 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-certs\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.848960 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-scripts\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.851709 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-config-data\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.859377 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lpmw\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-kube-api-access-7lpmw\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:40 crc kubenswrapper[4706]: I1208 19:46:40.862192 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-combined-ca-bundle\") pod \"cloudkitty-db-sync-wflqk\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:41 crc kubenswrapper[4706]: I1208 19:46:41.007107 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:46:41 crc kubenswrapper[4706]: I1208 19:46:41.658047 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e68efceb-8175-4600-bc86-0d9d5d9d5dbd" path="/var/lib/kubelet/pods/e68efceb-8175-4600-bc86-0d9d5d9d5dbd/volumes" Dec 08 19:46:41 crc kubenswrapper[4706]: I1208 19:46:41.708033 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-wflqk"] Dec 08 19:46:41 crc kubenswrapper[4706]: I1208 19:46:41.957511 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-wflqk" event={"ID":"a92db262-e832-42a3-8339-4f1824075e10","Type":"ContainerStarted","Data":"e87b82e2ed36789c6b3eec9a7d8d31b5d944a8a3f145786563f2fb5c8700667b"} Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.465879 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.786148 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.786543 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="ceilometer-central-agent" containerID="cri-o://a9b012f8e4bd28755c8343dd753f4ff3e89351c2c6f7f6f8c2f90a34226aebdd" gracePeriod=30 Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.786730 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="proxy-httpd" containerID="cri-o://6f01a8aeeb93a9681f3f65c2b065510158b3d67a4a1f6d3ac493ee976f665dbf" gracePeriod=30 Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.786769 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="sg-core" containerID="cri-o://1a66811b1cd05cf27bf4fe69f0869120c8952ee430cb1367f810ee022ac786b2" gracePeriod=30 Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.786800 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="ceilometer-notification-agent" containerID="cri-o://5ecbee74c514a01e6b24dd30a90a858b4eda5fe2013ac12f82d328d7764e6ab0" gracePeriod=30 Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.987556 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerID="6f01a8aeeb93a9681f3f65c2b065510158b3d67a4a1f6d3ac493ee976f665dbf" exitCode=0 Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.988097 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerID="1a66811b1cd05cf27bf4fe69f0869120c8952ee430cb1367f810ee022ac786b2" exitCode=2 Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.987629 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerDied","Data":"6f01a8aeeb93a9681f3f65c2b065510158b3d67a4a1f6d3ac493ee976f665dbf"} Dec 08 19:46:42 crc kubenswrapper[4706]: I1208 19:46:42.988178 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerDied","Data":"1a66811b1cd05cf27bf4fe69f0869120c8952ee430cb1367f810ee022ac786b2"} Dec 08 19:46:43 crc kubenswrapper[4706]: I1208 19:46:43.935034 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:46:44 crc kubenswrapper[4706]: I1208 19:46:44.024906 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerID="a9b012f8e4bd28755c8343dd753f4ff3e89351c2c6f7f6f8c2f90a34226aebdd" exitCode=0 Dec 08 19:46:44 crc kubenswrapper[4706]: I1208 19:46:44.024977 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerDied","Data":"a9b012f8e4bd28755c8343dd753f4ff3e89351c2c6f7f6f8c2f90a34226aebdd"} Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.104462 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerID="5ecbee74c514a01e6b24dd30a90a858b4eda5fe2013ac12f82d328d7764e6ab0" exitCode=0 Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.104679 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerDied","Data":"5ecbee74c514a01e6b24dd30a90a858b4eda5fe2013ac12f82d328d7764e6ab0"} Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.683236 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.758564 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-log-httpd\") pod \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.758669 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-scripts\") pod \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.758786 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-combined-ca-bundle\") pod \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.758878 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-sg-core-conf-yaml\") pod \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.758918 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-ceilometer-tls-certs\") pod \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.758951 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-config-data\") pod \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.759002 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p9gt\" (UniqueName: \"kubernetes.io/projected/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-kube-api-access-9p9gt\") pod \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.759091 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-run-httpd\") pod \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\" (UID: \"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3\") " Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.760465 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" (UID: "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.770175 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" (UID: "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.774972 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-scripts" (OuterVolumeSpecName: "scripts") pod "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" (UID: "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.801519 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-kube-api-access-9p9gt" (OuterVolumeSpecName: "kube-api-access-9p9gt") pod "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" (UID: "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3"). InnerVolumeSpecName "kube-api-access-9p9gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.857496 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" (UID: "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.870689 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.870738 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.870750 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.870758 4706 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.870776 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p9gt\" (UniqueName: \"kubernetes.io/projected/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-kube-api-access-9p9gt\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.914603 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" (UID: "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.974335 4706 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:47 crc kubenswrapper[4706]: I1208 19:46:47.989595 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" (UID: "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.017412 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-config-data" (OuterVolumeSpecName: "config-data") pod "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" (UID: "bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.077744 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.077789 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.162744 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3","Type":"ContainerDied","Data":"ff2daf3d218a3cfcd10424467fe53ef814f7ec7be2c6cdc3339c53f334cdf55e"} Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.162827 4706 scope.go:117] "RemoveContainer" containerID="6f01a8aeeb93a9681f3f65c2b065510158b3d67a4a1f6d3ac493ee976f665dbf" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.162995 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.197609 4706 scope.go:117] "RemoveContainer" containerID="1a66811b1cd05cf27bf4fe69f0869120c8952ee430cb1367f810ee022ac786b2" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.220416 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.232688 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.251498 4706 scope.go:117] "RemoveContainer" containerID="5ecbee74c514a01e6b24dd30a90a858b4eda5fe2013ac12f82d328d7764e6ab0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.261005 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:46:48 crc kubenswrapper[4706]: E1208 19:46:48.261680 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="ceilometer-central-agent" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.261708 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="ceilometer-central-agent" Dec 08 19:46:48 crc kubenswrapper[4706]: E1208 19:46:48.261729 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="sg-core" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.261739 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="sg-core" Dec 08 19:46:48 crc kubenswrapper[4706]: E1208 19:46:48.261756 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="proxy-httpd" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.261765 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="proxy-httpd" Dec 08 19:46:48 crc kubenswrapper[4706]: E1208 19:46:48.261806 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="ceilometer-notification-agent" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.261816 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="ceilometer-notification-agent" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.262113 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="ceilometer-notification-agent" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.262153 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="ceilometer-central-agent" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.262176 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="proxy-httpd" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.262195 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" containerName="sg-core" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.269229 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.275007 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.277767 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.278055 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.291003 4706 scope.go:117] "RemoveContainer" containerID="a9b012f8e4bd28755c8343dd753f4ff3e89351c2c6f7f6f8c2f90a34226aebdd" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.309171 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.409512 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-config-data\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.409574 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.409779 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-log-httpd\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.409848 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn9l9\" (UniqueName: \"kubernetes.io/projected/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-kube-api-access-xn9l9\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.409875 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-scripts\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.409909 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.409975 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.410092 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-run-httpd\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: E1208 19:46:48.475237 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeed8943_263a_449c_b219_72a8c0a838dd.slice/crio-conmon-97907e453fae5b40f9d718a1db50167194bd07aa926fc098fd20e8f74e473d8c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf8a8bb4_8e88_42ab_9d29_ffd0a21732b3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf8a8bb4_8e88_42ab_9d29_ffd0a21732b3.slice/crio-ff2daf3d218a3cfcd10424467fe53ef814f7ec7be2c6cdc3339c53f334cdf55e\": RecentStats: unable to find data in memory cache]" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.513781 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn9l9\" (UniqueName: \"kubernetes.io/projected/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-kube-api-access-xn9l9\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.513851 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-scripts\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.513892 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.513939 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.514043 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-run-httpd\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.514924 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-run-httpd\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.515819 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-config-data\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.515858 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.516015 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-log-httpd\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.516537 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-log-httpd\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.521561 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-scripts\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.521630 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.524165 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.525324 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.531712 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-config-data\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.538994 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn9l9\" (UniqueName: \"kubernetes.io/projected/e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42-kube-api-access-xn9l9\") pod \"ceilometer-0\" (UID: \"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42\") " pod="openstack/ceilometer-0" Dec 08 19:46:48 crc kubenswrapper[4706]: I1208 19:46:48.657820 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 19:46:49 crc kubenswrapper[4706]: I1208 19:46:49.016910 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerName="rabbitmq" containerID="cri-o://13404fedf045d4ea5efc3b98b3a810c9e6b8453164f7a0d6cc1b7e2ef389425e" gracePeriod=604794 Dec 08 19:46:49 crc kubenswrapper[4706]: I1208 19:46:49.300550 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 19:46:49 crc kubenswrapper[4706]: I1208 19:46:49.629868 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3" path="/var/lib/kubelet/pods/bf8a8bb4-8e88-42ab-9d29-ffd0a21732b3/volumes" Dec 08 19:46:50 crc kubenswrapper[4706]: I1208 19:46:50.203080 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42","Type":"ContainerStarted","Data":"7a2b0175f5a62a2ba523aa196e0a1c88652aaa2a2b9e1076b2e2026dcfa19868"} Dec 08 19:46:50 crc kubenswrapper[4706]: I1208 19:46:50.704004 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerName="rabbitmq" containerID="cri-o://3c8835b11dc1bed5c90dbdd9b422d6a651db9f230fb396efc8eaecee370eb7c3" gracePeriod=604794 Dec 08 19:46:55 crc kubenswrapper[4706]: I1208 19:46:55.187892 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Dec 08 19:46:55 crc kubenswrapper[4706]: I1208 19:46:55.680114 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.106:5671: connect: connection refused" Dec 08 19:46:56 crc kubenswrapper[4706]: I1208 19:46:56.310887 4706 generic.go:334] "Generic (PLEG): container finished" podID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerID="13404fedf045d4ea5efc3b98b3a810c9e6b8453164f7a0d6cc1b7e2ef389425e" exitCode=0 Dec 08 19:46:56 crc kubenswrapper[4706]: I1208 19:46:56.310991 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1472b8cb-795a-4a57-95f8-184a7707e8ad","Type":"ContainerDied","Data":"13404fedf045d4ea5efc3b98b3a810c9e6b8453164f7a0d6cc1b7e2ef389425e"} Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.348557 4706 generic.go:334] "Generic (PLEG): container finished" podID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerID="3c8835b11dc1bed5c90dbdd9b422d6a651db9f230fb396efc8eaecee370eb7c3" exitCode=0 Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.348656 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ea8337d-cff3-414c-98f6-17b41653b4fb","Type":"ContainerDied","Data":"3c8835b11dc1bed5c90dbdd9b422d6a651db9f230fb396efc8eaecee370eb7c3"} Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.601138 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-4pxb7"] Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.603790 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.609303 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.624245 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-4pxb7"] Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.758700 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-config\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.758856 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.758916 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.758948 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.759024 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdwbk\" (UniqueName: \"kubernetes.io/projected/8f25ea0e-071a-4399-96f7-0c41367a232f-kube-api-access-sdwbk\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.759075 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.759182 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.862138 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.862711 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.862741 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.862811 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdwbk\" (UniqueName: \"kubernetes.io/projected/8f25ea0e-071a-4399-96f7-0c41367a232f-kube-api-access-sdwbk\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.862863 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.862936 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.863170 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-config\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.864432 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.865384 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-config\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.865991 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.866451 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.866932 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.867358 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.891089 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdwbk\" (UniqueName: \"kubernetes.io/projected/8f25ea0e-071a-4399-96f7-0c41367a232f-kube-api-access-sdwbk\") pod \"dnsmasq-dns-dbb88bf8c-4pxb7\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:46:58 crc kubenswrapper[4706]: I1208 19:46:58.947562 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.682745 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745020 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-tls\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745085 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvb7f\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-kube-api-access-hvb7f\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745162 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1472b8cb-795a-4a57-95f8-184a7707e8ad-pod-info\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745229 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-plugins\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745311 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-confd\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745358 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-plugins-conf\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745457 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-config-data\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745503 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-erlang-cookie\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.745553 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1472b8cb-795a-4a57-95f8-184a7707e8ad-erlang-cookie-secret\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.746599 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.746634 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-server-conf\") pod \"1472b8cb-795a-4a57-95f8-184a7707e8ad\" (UID: \"1472b8cb-795a-4a57-95f8-184a7707e8ad\") " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.754853 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.755726 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.764170 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1472b8cb-795a-4a57-95f8-184a7707e8ad-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.770478 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.780424 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.791343 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1472b8cb-795a-4a57-95f8-184a7707e8ad-pod-info" (OuterVolumeSpecName: "pod-info") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.807913 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-kube-api-access-hvb7f" (OuterVolumeSpecName: "kube-api-access-hvb7f") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "kube-api-access-hvb7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.846095 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-config-data" (OuterVolumeSpecName: "config-data") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.853031 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c" (OuterVolumeSpecName: "persistence") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.856088 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.859828 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvb7f\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-kube-api-access-hvb7f\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.859923 4706 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1472b8cb-795a-4a57-95f8-184a7707e8ad-pod-info\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.860015 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.860084 4706 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.860155 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.860224 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.860376 4706 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1472b8cb-795a-4a57-95f8-184a7707e8ad-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.860485 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") on node \"crc\" " Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.956026 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.956253 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c") on node "crc" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.963645 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:01 crc kubenswrapper[4706]: I1208 19:47:01.964726 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-server-conf" (OuterVolumeSpecName: "server-conf") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.020428 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1472b8cb-795a-4a57-95f8-184a7707e8ad" (UID: "1472b8cb-795a-4a57-95f8-184a7707e8ad"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.065917 4706 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1472b8cb-795a-4a57-95f8-184a7707e8ad-server-conf\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.065988 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1472b8cb-795a-4a57-95f8-184a7707e8ad-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.427501 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1472b8cb-795a-4a57-95f8-184a7707e8ad","Type":"ContainerDied","Data":"d4e9fc451def0cba335a61435bed09c8b9d26cd0d42bfb2e536fff764f9ef677"} Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.428038 4706 scope.go:117] "RemoveContainer" containerID="13404fedf045d4ea5efc3b98b3a810c9e6b8453164f7a0d6cc1b7e2ef389425e" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.428829 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.499913 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.522923 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.546063 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:47:02 crc kubenswrapper[4706]: E1208 19:47:02.546950 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerName="rabbitmq" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.546972 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerName="rabbitmq" Dec 08 19:47:02 crc kubenswrapper[4706]: E1208 19:47:02.547019 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerName="setup-container" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.547027 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerName="setup-container" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.547396 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="1472b8cb-795a-4a57-95f8-184a7707e8ad" containerName="rabbitmq" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.549212 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.552347 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.552977 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.553982 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.554028 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.554521 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-cl2tg" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.554730 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.554886 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.566626 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.581463 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/faed709c-224b-4c30-9af3-d64d34525599-pod-info\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.581657 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc4bt\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-kube-api-access-sc4bt\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.581697 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.581757 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.581805 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-server-conf\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.582731 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.582840 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.582897 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.583063 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.583131 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-config-data\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.583516 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/faed709c-224b-4c30-9af3-d64d34525599-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.685981 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.686093 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.686135 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.686165 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-config-data\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.686188 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.686357 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/faed709c-224b-4c30-9af3-d64d34525599-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.686515 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/faed709c-224b-4c30-9af3-d64d34525599-pod-info\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.687744 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc4bt\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-kube-api-access-sc4bt\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.687814 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.688823 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.688916 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.689130 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.690255 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-config-data\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.691724 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/faed709c-224b-4c30-9af3-d64d34525599-pod-info\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.692052 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.692924 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-server-conf\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.693389 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.693453 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0cccd51a6a12e5e48b5c2f1604e3858ae97fd79f1ff6b61d3ff2b3945f701dee/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.694564 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/faed709c-224b-4c30-9af3-d64d34525599-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.695437 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/faed709c-224b-4c30-9af3-d64d34525599-server-conf\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.708752 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.711951 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.713233 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc4bt\" (UniqueName: \"kubernetes.io/projected/faed709c-224b-4c30-9af3-d64d34525599-kube-api-access-sc4bt\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.764252 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3ee3bd0-3e8d-4641-ba1f-2ab71f022a6c\") pod \"rabbitmq-server-0\" (UID: \"faed709c-224b-4c30-9af3-d64d34525599\") " pod="openstack/rabbitmq-server-0" Dec 08 19:47:02 crc kubenswrapper[4706]: I1208 19:47:02.949011 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 19:47:03 crc kubenswrapper[4706]: I1208 19:47:03.629364 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1472b8cb-795a-4a57-95f8-184a7707e8ad" path="/var/lib/kubelet/pods/1472b8cb-795a-4a57-95f8-184a7707e8ad/volumes" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.214800 4706 scope.go:117] "RemoveContainer" containerID="b9240b7d3b9c39a64fdc169ca933af330b5f84a9bba65f7ebdcde2ee2fe3e519" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.314171 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.377551 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-server-conf\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.377652 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-config-data\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.377697 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ea8337d-cff3-414c-98f6-17b41653b4fb-erlang-cookie-secret\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.377725 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-confd\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.380539 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.380603 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-tls\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.380691 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-plugins\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.380801 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ea8337d-cff3-414c-98f6-17b41653b4fb-pod-info\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.380862 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-erlang-cookie\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.380977 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-plugins-conf\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.381003 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhlbw\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-kube-api-access-rhlbw\") pod \"9ea8337d-cff3-414c-98f6-17b41653b4fb\" (UID: \"9ea8337d-cff3-414c-98f6-17b41653b4fb\") " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.389968 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.390427 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.390827 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.394330 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.394407 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9ea8337d-cff3-414c-98f6-17b41653b4fb-pod-info" (OuterVolumeSpecName: "pod-info") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.404823 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ea8337d-cff3-414c-98f6-17b41653b4fb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.404935 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-kube-api-access-rhlbw" (OuterVolumeSpecName: "kube-api-access-rhlbw") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "kube-api-access-rhlbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.450336 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37" (OuterVolumeSpecName: "persistence") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.485162 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-config-data" (OuterVolumeSpecName: "config-data") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506353 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506397 4706 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ea8337d-cff3-414c-98f6-17b41653b4fb-pod-info\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506409 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506424 4706 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506436 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhlbw\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-kube-api-access-rhlbw\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506447 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506457 4706 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ea8337d-cff3-414c-98f6-17b41653b4fb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506489 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") on node \"crc\" " Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.506500 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.536671 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ea8337d-cff3-414c-98f6-17b41653b4fb","Type":"ContainerDied","Data":"a615604c3637a30de2798fafe858b135087d5f0699a97e6ac7ed4daf394adbd4"} Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.536794 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.547147 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-server-conf" (OuterVolumeSpecName: "server-conf") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.581907 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.582272 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37") on node "crc" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.610739 4706 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ea8337d-cff3-414c-98f6-17b41653b4fb-server-conf\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.610787 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.672117 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9ea8337d-cff3-414c-98f6-17b41653b4fb" (UID: "9ea8337d-cff3-414c-98f6-17b41653b4fb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.713733 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ea8337d-cff3-414c-98f6-17b41653b4fb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.899730 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.926018 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.946965 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:47:05 crc kubenswrapper[4706]: E1208 19:47:05.947677 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerName="rabbitmq" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.947700 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerName="rabbitmq" Dec 08 19:47:05 crc kubenswrapper[4706]: E1208 19:47:05.947718 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerName="setup-container" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.947727 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerName="setup-container" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.948038 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ea8337d-cff3-414c-98f6-17b41653b4fb" containerName="rabbitmq" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.949841 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.958044 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.958349 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.958522 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.958608 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-glhtm" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.958618 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.958762 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.958675 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 08 19:47:05 crc kubenswrapper[4706]: I1208 19:47:05.976613 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.023713 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtv5w\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-kube-api-access-qtv5w\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.023799 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.023848 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.023886 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.023922 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.024165 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.024367 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.024461 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.024724 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.024896 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.024960 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.126869 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.126927 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.126964 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127002 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127031 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127454 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127648 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127562 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127730 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127827 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127876 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.127978 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtv5w\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-kube-api-access-qtv5w\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.128032 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.129048 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.129196 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.129294 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.131643 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.131675 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8f2cec159fcbaca32539204793034491d123d827594a614a5fcc96cb10f96558/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.135662 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.135992 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.137688 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.137701 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.155819 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtv5w\" (UniqueName: \"kubernetes.io/projected/ecc9ad61-4d9c-4d26-a240-398cfc8c3165-kube-api-access-qtv5w\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.184039 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9de65bc-6cb0-4c1c-96e9-e599376fbc37\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecc9ad61-4d9c-4d26-a240-398cfc8c3165\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: I1208 19:47:06.293411 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:06 crc kubenswrapper[4706]: E1208 19:47:06.387284 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current-tested" Dec 08 19:47:06 crc kubenswrapper[4706]: E1208 19:47:06.387345 4706 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current-tested" Dec 08 19:47:06 crc kubenswrapper[4706]: E1208 19:47:06.387503 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7lpmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42406,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-wflqk_openstack(a92db262-e832-42a3-8339-4f1824075e10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 19:47:06 crc kubenswrapper[4706]: E1208 19:47:06.388993 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-wflqk" podUID="a92db262-e832-42a3-8339-4f1824075e10" Dec 08 19:47:06 crc kubenswrapper[4706]: E1208 19:47:06.558981 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current-tested\\\"\"" pod="openstack/cloudkitty-db-sync-wflqk" podUID="a92db262-e832-42a3-8339-4f1824075e10" Dec 08 19:47:07 crc kubenswrapper[4706]: I1208 19:47:07.658477 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ea8337d-cff3-414c-98f6-17b41653b4fb" path="/var/lib/kubelet/pods/9ea8337d-cff3-414c-98f6-17b41653b4fb/volumes" Dec 08 19:47:08 crc kubenswrapper[4706]: I1208 19:47:08.268302 4706 scope.go:117] "RemoveContainer" containerID="3c8835b11dc1bed5c90dbdd9b422d6a651db9f230fb396efc8eaecee370eb7c3" Dec 08 19:47:08 crc kubenswrapper[4706]: I1208 19:47:08.354046 4706 scope.go:117] "RemoveContainer" containerID="4d2558eb0a6a323f3921e28819b25ee1b7d270e9ab07768a3bfcedfd8570cdbd" Dec 08 19:47:08 crc kubenswrapper[4706]: I1208 19:47:08.825123 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-4pxb7"] Dec 08 19:47:08 crc kubenswrapper[4706]: I1208 19:47:08.977674 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 19:47:09 crc kubenswrapper[4706]: I1208 19:47:09.002988 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 19:47:09 crc kubenswrapper[4706]: W1208 19:47:09.022034 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecc9ad61_4d9c_4d26_a240_398cfc8c3165.slice/crio-2f1f0786bebc511baa389fd1444a01df32a5a3d006bb208fb7ec4d8f324b37f7 WatchSource:0}: Error finding container 2f1f0786bebc511baa389fd1444a01df32a5a3d006bb208fb7ec4d8f324b37f7: Status 404 returned error can't find the container with id 2f1f0786bebc511baa389fd1444a01df32a5a3d006bb208fb7ec4d8f324b37f7 Dec 08 19:47:09 crc kubenswrapper[4706]: I1208 19:47:09.607098 4706 generic.go:334] "Generic (PLEG): container finished" podID="8f25ea0e-071a-4399-96f7-0c41367a232f" containerID="a758e87f1f4823ed673d33798d82f9b711d83da401be85ecf30a8f13a5323774" exitCode=0 Dec 08 19:47:09 crc kubenswrapper[4706]: I1208 19:47:09.607228 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" event={"ID":"8f25ea0e-071a-4399-96f7-0c41367a232f","Type":"ContainerDied","Data":"a758e87f1f4823ed673d33798d82f9b711d83da401be85ecf30a8f13a5323774"} Dec 08 19:47:09 crc kubenswrapper[4706]: I1208 19:47:09.608144 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" event={"ID":"8f25ea0e-071a-4399-96f7-0c41367a232f","Type":"ContainerStarted","Data":"6cff4f033810a27e6304bde3c50d6cb6d53186ffa1ecdae8b9529b78911d337d"} Dec 08 19:47:09 crc kubenswrapper[4706]: I1208 19:47:09.627053 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42","Type":"ContainerStarted","Data":"f4f4c18082e1e1faffd29ed84b2f862044e09c692698bd3e1f745167c4fa3066"} Dec 08 19:47:09 crc kubenswrapper[4706]: I1208 19:47:09.627136 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecc9ad61-4d9c-4d26-a240-398cfc8c3165","Type":"ContainerStarted","Data":"2f1f0786bebc511baa389fd1444a01df32a5a3d006bb208fb7ec4d8f324b37f7"} Dec 08 19:47:09 crc kubenswrapper[4706]: I1208 19:47:09.627152 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"faed709c-224b-4c30-9af3-d64d34525599","Type":"ContainerStarted","Data":"6f817d68d7d4ac5f13900913f57c625825e741437818c0df51ca3b69ed6c40e4"} Dec 08 19:47:10 crc kubenswrapper[4706]: I1208 19:47:10.634221 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42","Type":"ContainerStarted","Data":"daa6e748e98a9d5002b8319c5308e2dc8a7b0038ee496ad02336d2c374a934bb"} Dec 08 19:47:10 crc kubenswrapper[4706]: I1208 19:47:10.637799 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" event={"ID":"8f25ea0e-071a-4399-96f7-0c41367a232f","Type":"ContainerStarted","Data":"ecb91bb37eeb4ff49424a5ed7f41da4468c7747d96e42b4f9f92207894d1af91"} Dec 08 19:47:10 crc kubenswrapper[4706]: I1208 19:47:10.637957 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:47:10 crc kubenswrapper[4706]: I1208 19:47:10.678050 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" podStartSLOduration=12.678019792 podStartE2EDuration="12.678019792s" podCreationTimestamp="2025-12-08 19:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:47:10.660917755 +0000 UTC m=+1513.303118768" watchObservedRunningTime="2025-12-08 19:47:10.678019792 +0000 UTC m=+1513.320220815" Dec 08 19:47:12 crc kubenswrapper[4706]: I1208 19:47:12.670085 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42","Type":"ContainerStarted","Data":"3c5002ab3f173b77ad68d07929f64c6125ee6cb16997e48dade4eadabc36dfa2"} Dec 08 19:47:12 crc kubenswrapper[4706]: I1208 19:47:12.675553 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecc9ad61-4d9c-4d26-a240-398cfc8c3165","Type":"ContainerStarted","Data":"1bbcfb38eef75278ddd8dcd66ef17d433522e1fdce4e1162896c39cbc0ac2fdb"} Dec 08 19:47:12 crc kubenswrapper[4706]: I1208 19:47:12.678935 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"faed709c-224b-4c30-9af3-d64d34525599","Type":"ContainerStarted","Data":"f8ea3e55f55046de20a740b5a047ac905eeae4750bb017877973debfa7058da0"} Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.675830 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fz5kt"] Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.681822 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.711797 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fz5kt"] Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.862923 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clsp9\" (UniqueName: \"kubernetes.io/projected/dd14ee90-6391-469b-b808-0a862f8683a5-kube-api-access-clsp9\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.863065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-catalog-content\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.863206 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-utilities\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.965577 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-utilities\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.965755 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clsp9\" (UniqueName: \"kubernetes.io/projected/dd14ee90-6391-469b-b808-0a862f8683a5-kube-api-access-clsp9\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.965847 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-catalog-content\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.966129 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-utilities\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.973479 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-catalog-content\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:13 crc kubenswrapper[4706]: I1208 19:47:13.996935 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clsp9\" (UniqueName: \"kubernetes.io/projected/dd14ee90-6391-469b-b808-0a862f8683a5-kube-api-access-clsp9\") pod \"certified-operators-fz5kt\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:14 crc kubenswrapper[4706]: I1208 19:47:14.079410 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:15 crc kubenswrapper[4706]: I1208 19:47:14.591733 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fz5kt"] Dec 08 19:47:15 crc kubenswrapper[4706]: W1208 19:47:14.593821 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd14ee90_6391_469b_b808_0a862f8683a5.slice/crio-8c2e77bdf67f3f8cd6d41322f5368aa9e680e274c058f74c49b4017defbd055e WatchSource:0}: Error finding container 8c2e77bdf67f3f8cd6d41322f5368aa9e680e274c058f74c49b4017defbd055e: Status 404 returned error can't find the container with id 8c2e77bdf67f3f8cd6d41322f5368aa9e680e274c058f74c49b4017defbd055e Dec 08 19:47:15 crc kubenswrapper[4706]: I1208 19:47:14.736066 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fz5kt" event={"ID":"dd14ee90-6391-469b-b808-0a862f8683a5","Type":"ContainerStarted","Data":"8c2e77bdf67f3f8cd6d41322f5368aa9e680e274c058f74c49b4017defbd055e"} Dec 08 19:47:15 crc kubenswrapper[4706]: I1208 19:47:14.739773 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42","Type":"ContainerStarted","Data":"cba305ac60224802553e2f0007f1aa7a13da9400638ca4cf8878ed473babbbd0"} Dec 08 19:47:15 crc kubenswrapper[4706]: I1208 19:47:14.741803 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 19:47:15 crc kubenswrapper[4706]: I1208 19:47:14.772900 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.54324381 podStartE2EDuration="26.772875633s" podCreationTimestamp="2025-12-08 19:46:48 +0000 UTC" firstStartedPulling="2025-12-08 19:46:49.321907665 +0000 UTC m=+1491.964108668" lastFinishedPulling="2025-12-08 19:47:13.551539498 +0000 UTC m=+1516.193740491" observedRunningTime="2025-12-08 19:47:14.766578454 +0000 UTC m=+1517.408779457" watchObservedRunningTime="2025-12-08 19:47:14.772875633 +0000 UTC m=+1517.415076636" Dec 08 19:47:15 crc kubenswrapper[4706]: I1208 19:47:15.755027 4706 generic.go:334] "Generic (PLEG): container finished" podID="dd14ee90-6391-469b-b808-0a862f8683a5" containerID="7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510" exitCode=0 Dec 08 19:47:15 crc kubenswrapper[4706]: I1208 19:47:15.755354 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fz5kt" event={"ID":"dd14ee90-6391-469b-b808-0a862f8683a5","Type":"ContainerDied","Data":"7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510"} Dec 08 19:47:16 crc kubenswrapper[4706]: I1208 19:47:16.788315 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fz5kt" event={"ID":"dd14ee90-6391-469b-b808-0a862f8683a5","Type":"ContainerStarted","Data":"2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9"} Dec 08 19:47:18 crc kubenswrapper[4706]: I1208 19:47:18.829862 4706 generic.go:334] "Generic (PLEG): container finished" podID="dd14ee90-6391-469b-b808-0a862f8683a5" containerID="2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9" exitCode=0 Dec 08 19:47:18 crc kubenswrapper[4706]: I1208 19:47:18.829991 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fz5kt" event={"ID":"dd14ee90-6391-469b-b808-0a862f8683a5","Type":"ContainerDied","Data":"2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9"} Dec 08 19:47:18 crc kubenswrapper[4706]: I1208 19:47:18.949582 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.014935 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-4bzh4"] Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.015245 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" podUID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" containerName="dnsmasq-dns" containerID="cri-o://8f4b403bc6c974c5ff8255c185cc2a11665f2efee113e234d62ef32b8d0a3a2e" gracePeriod=10 Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.249543 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.315551 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-862hf"] Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.318708 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.361600 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-862hf"] Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.414390 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-config\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.414447 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.414479 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.414509 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-dns-svc\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.414534 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jgk4\" (UniqueName: \"kubernetes.io/projected/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-kube-api-access-6jgk4\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.414604 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.414637 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.521197 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-config\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.521284 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.521320 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.521366 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-dns-svc\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.521398 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jgk4\" (UniqueName: \"kubernetes.io/projected/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-kube-api-access-6jgk4\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.521494 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.521540 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.522598 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.523136 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-config\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.524386 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.525183 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-dns-svc\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.525969 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.527154 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.552602 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jgk4\" (UniqueName: \"kubernetes.io/projected/797f5d64-d90a-4466-95f7-a9b4b43d9f1c-kube-api-access-6jgk4\") pod \"dnsmasq-dns-85f64749dc-862hf\" (UID: \"797f5d64-d90a-4466-95f7-a9b4b43d9f1c\") " pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.783902 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.862853 4706 generic.go:334] "Generic (PLEG): container finished" podID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" containerID="8f4b403bc6c974c5ff8255c185cc2a11665f2efee113e234d62ef32b8d0a3a2e" exitCode=0 Dec 08 19:47:19 crc kubenswrapper[4706]: I1208 19:47:19.862963 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" event={"ID":"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa","Type":"ContainerDied","Data":"8f4b403bc6c974c5ff8255c185cc2a11665f2efee113e234d62ef32b8d0a3a2e"} Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.267943 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.455293 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-sb\") pod \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.455429 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrthz\" (UniqueName: \"kubernetes.io/projected/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-kube-api-access-mrthz\") pod \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.455503 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-svc\") pod \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.455577 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-nb\") pod \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.455647 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config\") pod \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.455695 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-swift-storage-0\") pod \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.467543 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-862hf"] Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.496134 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-kube-api-access-mrthz" (OuterVolumeSpecName: "kube-api-access-mrthz") pod "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" (UID: "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa"). InnerVolumeSpecName "kube-api-access-mrthz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.551887 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" (UID: "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.559247 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.559299 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrthz\" (UniqueName: \"kubernetes.io/projected/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-kube-api-access-mrthz\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.563126 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" (UID: "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.588315 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" (UID: "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:20 crc kubenswrapper[4706]: E1208 19:47:20.597591 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config podName:2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa nodeName:}" failed. No retries permitted until 2025-12-08 19:47:21.097548392 +0000 UTC m=+1523.739749395 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config") pod "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" (UID: "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa") : error deleting /var/lib/kubelet/pods/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa/volume-subpaths: remove /var/lib/kubelet/pods/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa/volume-subpaths: no such file or directory Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.598085 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" (UID: "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.662072 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.662539 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.662555 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.882362 4706 generic.go:334] "Generic (PLEG): container finished" podID="797f5d64-d90a-4466-95f7-a9b4b43d9f1c" containerID="7236ca8e352a01c4f3c0d76f1af4c1680b2b296ca8fe9767abc7db624366c1ef" exitCode=0 Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.882422 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-862hf" event={"ID":"797f5d64-d90a-4466-95f7-a9b4b43d9f1c","Type":"ContainerDied","Data":"7236ca8e352a01c4f3c0d76f1af4c1680b2b296ca8fe9767abc7db624366c1ef"} Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.882462 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-862hf" event={"ID":"797f5d64-d90a-4466-95f7-a9b4b43d9f1c","Type":"ContainerStarted","Data":"add8f959926761a7793d53d69d76c2c0d1ae8dde780a5910775ef89ffe667e0a"} Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.885952 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fz5kt" event={"ID":"dd14ee90-6391-469b-b808-0a862f8683a5","Type":"ContainerStarted","Data":"320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44"} Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.889202 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" event={"ID":"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa","Type":"ContainerDied","Data":"fe888104256b4b207cc2e4846ac8485159f732418062e2aee9be7e4404178404"} Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.889273 4706 scope.go:117] "RemoveContainer" containerID="8f4b403bc6c974c5ff8255c185cc2a11665f2efee113e234d62ef32b8d0a3a2e" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.889478 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-4bzh4" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.893636 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-wflqk" event={"ID":"a92db262-e832-42a3-8339-4f1824075e10","Type":"ContainerStarted","Data":"94bcd813bba47a429faf7a12e2270b23d2300b5b7e5e233de6ee69fc552c9097"} Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.932861 4706 scope.go:117] "RemoveContainer" containerID="4f1206e835725090ac898784cb6a41086707eb82045d9a4b8d17de88f475ab04" Dec 08 19:47:20 crc kubenswrapper[4706]: I1208 19:47:20.968500 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fz5kt" podStartSLOduration=4.26882697 podStartE2EDuration="7.968476351s" podCreationTimestamp="2025-12-08 19:47:13 +0000 UTC" firstStartedPulling="2025-12-08 19:47:15.757922409 +0000 UTC m=+1518.400123412" lastFinishedPulling="2025-12-08 19:47:19.45757179 +0000 UTC m=+1522.099772793" observedRunningTime="2025-12-08 19:47:20.945458956 +0000 UTC m=+1523.587659959" watchObservedRunningTime="2025-12-08 19:47:20.968476351 +0000 UTC m=+1523.610677354" Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.009744 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-wflqk" podStartSLOduration=3.479969721 podStartE2EDuration="41.009721564s" podCreationTimestamp="2025-12-08 19:46:40 +0000 UTC" firstStartedPulling="2025-12-08 19:46:41.715832578 +0000 UTC m=+1484.358033581" lastFinishedPulling="2025-12-08 19:47:19.245584421 +0000 UTC m=+1521.887785424" observedRunningTime="2025-12-08 19:47:20.974643256 +0000 UTC m=+1523.616844259" watchObservedRunningTime="2025-12-08 19:47:21.009721564 +0000 UTC m=+1523.651922567" Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.175416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config\") pod \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\" (UID: \"2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa\") " Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.177578 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config" (OuterVolumeSpecName: "config") pod "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" (UID: "2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.240682 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-4bzh4"] Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.249987 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-4bzh4"] Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.279580 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.621403 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" path="/var/lib/kubelet/pods/2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa/volumes" Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.913006 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-862hf" event={"ID":"797f5d64-d90a-4466-95f7-a9b4b43d9f1c","Type":"ContainerStarted","Data":"d392388597eb1f331813f7220eb341a9f86748aa0f7ebc8fbec84499b95d09d2"} Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.913492 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:21 crc kubenswrapper[4706]: I1208 19:47:21.937155 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85f64749dc-862hf" podStartSLOduration=2.93713536 podStartE2EDuration="2.93713536s" podCreationTimestamp="2025-12-08 19:47:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:47:21.934487575 +0000 UTC m=+1524.576688598" watchObservedRunningTime="2025-12-08 19:47:21.93713536 +0000 UTC m=+1524.579336373" Dec 08 19:47:22 crc kubenswrapper[4706]: I1208 19:47:22.926356 4706 generic.go:334] "Generic (PLEG): container finished" podID="a92db262-e832-42a3-8339-4f1824075e10" containerID="94bcd813bba47a429faf7a12e2270b23d2300b5b7e5e233de6ee69fc552c9097" exitCode=0 Dec 08 19:47:22 crc kubenswrapper[4706]: I1208 19:47:22.926446 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-wflqk" event={"ID":"a92db262-e832-42a3-8339-4f1824075e10","Type":"ContainerDied","Data":"94bcd813bba47a429faf7a12e2270b23d2300b5b7e5e233de6ee69fc552c9097"} Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.080804 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.081175 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.141454 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.381306 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.557103 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-scripts\") pod \"a92db262-e832-42a3-8339-4f1824075e10\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.558587 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-combined-ca-bundle\") pod \"a92db262-e832-42a3-8339-4f1824075e10\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.558687 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-certs\") pod \"a92db262-e832-42a3-8339-4f1824075e10\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.558822 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lpmw\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-kube-api-access-7lpmw\") pod \"a92db262-e832-42a3-8339-4f1824075e10\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.558964 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-config-data\") pod \"a92db262-e832-42a3-8339-4f1824075e10\" (UID: \"a92db262-e832-42a3-8339-4f1824075e10\") " Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.566563 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-kube-api-access-7lpmw" (OuterVolumeSpecName: "kube-api-access-7lpmw") pod "a92db262-e832-42a3-8339-4f1824075e10" (UID: "a92db262-e832-42a3-8339-4f1824075e10"). InnerVolumeSpecName "kube-api-access-7lpmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.566631 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-certs" (OuterVolumeSpecName: "certs") pod "a92db262-e832-42a3-8339-4f1824075e10" (UID: "a92db262-e832-42a3-8339-4f1824075e10"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.567107 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-scripts" (OuterVolumeSpecName: "scripts") pod "a92db262-e832-42a3-8339-4f1824075e10" (UID: "a92db262-e832-42a3-8339-4f1824075e10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.592361 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-config-data" (OuterVolumeSpecName: "config-data") pod "a92db262-e832-42a3-8339-4f1824075e10" (UID: "a92db262-e832-42a3-8339-4f1824075e10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.595668 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a92db262-e832-42a3-8339-4f1824075e10" (UID: "a92db262-e832-42a3-8339-4f1824075e10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.663965 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.664012 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.664023 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92db262-e832-42a3-8339-4f1824075e10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.664037 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.664053 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lpmw\" (UniqueName: \"kubernetes.io/projected/a92db262-e832-42a3-8339-4f1824075e10-kube-api-access-7lpmw\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.953752 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-wflqk" Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.953886 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-wflqk" event={"ID":"a92db262-e832-42a3-8339-4f1824075e10","Type":"ContainerDied","Data":"e87b82e2ed36789c6b3eec9a7d8d31b5d944a8a3f145786563f2fb5c8700667b"} Dec 08 19:47:24 crc kubenswrapper[4706]: I1208 19:47:24.954794 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e87b82e2ed36789c6b3eec9a7d8d31b5d944a8a3f145786563f2fb5c8700667b" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.028111 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.099582 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fz5kt"] Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.496620 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-np9k8"] Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.512012 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-np9k8"] Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.601768 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-j2q44"] Dec 08 19:47:25 crc kubenswrapper[4706]: E1208 19:47:25.602445 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a92db262-e832-42a3-8339-4f1824075e10" containerName="cloudkitty-db-sync" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.602465 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a92db262-e832-42a3-8339-4f1824075e10" containerName="cloudkitty-db-sync" Dec 08 19:47:25 crc kubenswrapper[4706]: E1208 19:47:25.602505 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" containerName="dnsmasq-dns" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.602514 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" containerName="dnsmasq-dns" Dec 08 19:47:25 crc kubenswrapper[4706]: E1208 19:47:25.603469 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" containerName="init" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.603493 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" containerName="init" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.603750 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a92db262-e832-42a3-8339-4f1824075e10" containerName="cloudkitty-db-sync" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.603777 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bb3bbc2-c41f-4441-972d-7ac6fe2abfaa" containerName="dnsmasq-dns" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.605008 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.608292 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.633887 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1" path="/var/lib/kubelet/pods/2bc9f1ec-bca2-4f07-bdf9-d91d523ccaf1/volumes" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.634937 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-j2q44"] Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.689958 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-certs\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.690023 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-config-data\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.690148 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt7mf\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-kube-api-access-bt7mf\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.690185 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-combined-ca-bundle\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.690272 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-scripts\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.794696 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt7mf\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-kube-api-access-bt7mf\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.794879 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-combined-ca-bundle\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.795298 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-scripts\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.795513 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-certs\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.795611 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-config-data\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.800905 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-combined-ca-bundle\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.803441 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-certs\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.817164 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt7mf\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-kube-api-access-bt7mf\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.818168 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-config-data\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.838633 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-scripts\") pod \"cloudkitty-storageinit-j2q44\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:25 crc kubenswrapper[4706]: I1208 19:47:25.932402 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:26 crc kubenswrapper[4706]: I1208 19:47:26.466031 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-j2q44"] Dec 08 19:47:26 crc kubenswrapper[4706]: W1208 19:47:26.472369 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98651db7_0458_47ab_8621_7098bf5337ae.slice/crio-4b460e45fc9bfe657f1c5945ab8f3de238db224c601a0210797fe7f3503bf975 WatchSource:0}: Error finding container 4b460e45fc9bfe657f1c5945ab8f3de238db224c601a0210797fe7f3503bf975: Status 404 returned error can't find the container with id 4b460e45fc9bfe657f1c5945ab8f3de238db224c601a0210797fe7f3503bf975 Dec 08 19:47:26 crc kubenswrapper[4706]: I1208 19:47:26.976585 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-j2q44" event={"ID":"98651db7-0458-47ab-8621-7098bf5337ae","Type":"ContainerStarted","Data":"e0ceaabd43655b7fa5356093f0759d3fbfe41f5fec8ec7421e65583bc11a0d44"} Dec 08 19:47:26 crc kubenswrapper[4706]: I1208 19:47:26.977002 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-j2q44" event={"ID":"98651db7-0458-47ab-8621-7098bf5337ae","Type":"ContainerStarted","Data":"4b460e45fc9bfe657f1c5945ab8f3de238db224c601a0210797fe7f3503bf975"} Dec 08 19:47:26 crc kubenswrapper[4706]: I1208 19:47:26.977367 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fz5kt" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" containerName="registry-server" containerID="cri-o://320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44" gracePeriod=2 Dec 08 19:47:26 crc kubenswrapper[4706]: I1208 19:47:26.999060 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-j2q44" podStartSLOduration=1.999034946 podStartE2EDuration="1.999034946s" podCreationTimestamp="2025-12-08 19:47:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:47:26.993747396 +0000 UTC m=+1529.635948399" watchObservedRunningTime="2025-12-08 19:47:26.999034946 +0000 UTC m=+1529.641235949" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.680009 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.858735 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-catalog-content\") pod \"dd14ee90-6391-469b-b808-0a862f8683a5\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.858913 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clsp9\" (UniqueName: \"kubernetes.io/projected/dd14ee90-6391-469b-b808-0a862f8683a5-kube-api-access-clsp9\") pod \"dd14ee90-6391-469b-b808-0a862f8683a5\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.858949 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-utilities\") pod \"dd14ee90-6391-469b-b808-0a862f8683a5\" (UID: \"dd14ee90-6391-469b-b808-0a862f8683a5\") " Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.860951 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-utilities" (OuterVolumeSpecName: "utilities") pod "dd14ee90-6391-469b-b808-0a862f8683a5" (UID: "dd14ee90-6391-469b-b808-0a862f8683a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.869865 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd14ee90-6391-469b-b808-0a862f8683a5-kube-api-access-clsp9" (OuterVolumeSpecName: "kube-api-access-clsp9") pod "dd14ee90-6391-469b-b808-0a862f8683a5" (UID: "dd14ee90-6391-469b-b808-0a862f8683a5"). InnerVolumeSpecName "kube-api-access-clsp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.954422 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd14ee90-6391-469b-b808-0a862f8683a5" (UID: "dd14ee90-6391-469b-b808-0a862f8683a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.962067 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.962104 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clsp9\" (UniqueName: \"kubernetes.io/projected/dd14ee90-6391-469b-b808-0a862f8683a5-kube-api-access-clsp9\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.962117 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd14ee90-6391-469b-b808-0a862f8683a5-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.993535 4706 generic.go:334] "Generic (PLEG): container finished" podID="dd14ee90-6391-469b-b808-0a862f8683a5" containerID="320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44" exitCode=0 Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.993986 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fz5kt" event={"ID":"dd14ee90-6391-469b-b808-0a862f8683a5","Type":"ContainerDied","Data":"320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44"} Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.994035 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fz5kt" Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.994071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fz5kt" event={"ID":"dd14ee90-6391-469b-b808-0a862f8683a5","Type":"ContainerDied","Data":"8c2e77bdf67f3f8cd6d41322f5368aa9e680e274c058f74c49b4017defbd055e"} Dec 08 19:47:27 crc kubenswrapper[4706]: I1208 19:47:27.994095 4706 scope.go:117] "RemoveContainer" containerID="320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44" Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.048231 4706 scope.go:117] "RemoveContainer" containerID="2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9" Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.060746 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fz5kt"] Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.075324 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fz5kt"] Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.083429 4706 scope.go:117] "RemoveContainer" containerID="7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510" Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.140458 4706 scope.go:117] "RemoveContainer" containerID="320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44" Dec 08 19:47:28 crc kubenswrapper[4706]: E1208 19:47:28.142006 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44\": container with ID starting with 320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44 not found: ID does not exist" containerID="320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44" Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.142103 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44"} err="failed to get container status \"320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44\": rpc error: code = NotFound desc = could not find container \"320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44\": container with ID starting with 320d955d972a410f3d43e267b20a0f66c33758313b48387200b2932bc21c7b44 not found: ID does not exist" Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.142171 4706 scope.go:117] "RemoveContainer" containerID="2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9" Dec 08 19:47:28 crc kubenswrapper[4706]: E1208 19:47:28.142876 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9\": container with ID starting with 2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9 not found: ID does not exist" containerID="2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9" Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.142923 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9"} err="failed to get container status \"2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9\": rpc error: code = NotFound desc = could not find container \"2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9\": container with ID starting with 2fa01414414725653c8deb67bf70f726cf229de3eaa6c229a7303e8d9f7406d9 not found: ID does not exist" Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.143015 4706 scope.go:117] "RemoveContainer" containerID="7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510" Dec 08 19:47:28 crc kubenswrapper[4706]: E1208 19:47:28.143637 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510\": container with ID starting with 7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510 not found: ID does not exist" containerID="7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510" Dec 08 19:47:28 crc kubenswrapper[4706]: I1208 19:47:28.143665 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510"} err="failed to get container status \"7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510\": rpc error: code = NotFound desc = could not find container \"7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510\": container with ID starting with 7148ec9096af2c39ef6d10611c7bfe7b89b47f323622e3345314b5f778bc5510 not found: ID does not exist" Dec 08 19:47:29 crc kubenswrapper[4706]: I1208 19:47:29.008091 4706 generic.go:334] "Generic (PLEG): container finished" podID="98651db7-0458-47ab-8621-7098bf5337ae" containerID="e0ceaabd43655b7fa5356093f0759d3fbfe41f5fec8ec7421e65583bc11a0d44" exitCode=0 Dec 08 19:47:29 crc kubenswrapper[4706]: I1208 19:47:29.008134 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-j2q44" event={"ID":"98651db7-0458-47ab-8621-7098bf5337ae","Type":"ContainerDied","Data":"e0ceaabd43655b7fa5356093f0759d3fbfe41f5fec8ec7421e65583bc11a0d44"} Dec 08 19:47:29 crc kubenswrapper[4706]: I1208 19:47:29.621136 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" path="/var/lib/kubelet/pods/dd14ee90-6391-469b-b808-0a862f8683a5/volumes" Dec 08 19:47:29 crc kubenswrapper[4706]: I1208 19:47:29.785982 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85f64749dc-862hf" Dec 08 19:47:29 crc kubenswrapper[4706]: I1208 19:47:29.883182 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-4pxb7"] Dec 08 19:47:29 crc kubenswrapper[4706]: I1208 19:47:29.884461 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" podUID="8f25ea0e-071a-4399-96f7-0c41367a232f" containerName="dnsmasq-dns" containerID="cri-o://ecb91bb37eeb4ff49424a5ed7f41da4468c7747d96e42b4f9f92207894d1af91" gracePeriod=10 Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.024000 4706 generic.go:334] "Generic (PLEG): container finished" podID="8f25ea0e-071a-4399-96f7-0c41367a232f" containerID="ecb91bb37eeb4ff49424a5ed7f41da4468c7747d96e42b4f9f92207894d1af91" exitCode=0 Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.024104 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" event={"ID":"8f25ea0e-071a-4399-96f7-0c41367a232f","Type":"ContainerDied","Data":"ecb91bb37eeb4ff49424a5ed7f41da4468c7747d96e42b4f9f92207894d1af91"} Dec 08 19:47:30 crc kubenswrapper[4706]: E1208 19:47:30.103279 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f25ea0e_071a_4399_96f7_0c41367a232f.slice/crio-ecb91bb37eeb4ff49424a5ed7f41da4468c7747d96e42b4f9f92207894d1af91.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f25ea0e_071a_4399_96f7_0c41367a232f.slice/crio-conmon-ecb91bb37eeb4ff49424a5ed7f41da4468c7747d96e42b4f9f92207894d1af91.scope\": RecentStats: unable to find data in memory cache]" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.570491 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.744519 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-certs\") pod \"98651db7-0458-47ab-8621-7098bf5337ae\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.744592 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-config-data\") pod \"98651db7-0458-47ab-8621-7098bf5337ae\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.744753 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-scripts\") pod \"98651db7-0458-47ab-8621-7098bf5337ae\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.744864 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt7mf\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-kube-api-access-bt7mf\") pod \"98651db7-0458-47ab-8621-7098bf5337ae\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.744900 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-combined-ca-bundle\") pod \"98651db7-0458-47ab-8621-7098bf5337ae\" (UID: \"98651db7-0458-47ab-8621-7098bf5337ae\") " Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.756570 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-certs" (OuterVolumeSpecName: "certs") pod "98651db7-0458-47ab-8621-7098bf5337ae" (UID: "98651db7-0458-47ab-8621-7098bf5337ae"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.756718 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-scripts" (OuterVolumeSpecName: "scripts") pod "98651db7-0458-47ab-8621-7098bf5337ae" (UID: "98651db7-0458-47ab-8621-7098bf5337ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.767605 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-kube-api-access-bt7mf" (OuterVolumeSpecName: "kube-api-access-bt7mf") pod "98651db7-0458-47ab-8621-7098bf5337ae" (UID: "98651db7-0458-47ab-8621-7098bf5337ae"). InnerVolumeSpecName "kube-api-access-bt7mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.814543 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98651db7-0458-47ab-8621-7098bf5337ae" (UID: "98651db7-0458-47ab-8621-7098bf5337ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.814725 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-config-data" (OuterVolumeSpecName: "config-data") pod "98651db7-0458-47ab-8621-7098bf5337ae" (UID: "98651db7-0458-47ab-8621-7098bf5337ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.848035 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.848085 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.848096 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.848104 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98651db7-0458-47ab-8621-7098bf5337ae-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.848115 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt7mf\" (UniqueName: \"kubernetes.io/projected/98651db7-0458-47ab-8621-7098bf5337ae-kube-api-access-bt7mf\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:30 crc kubenswrapper[4706]: I1208 19:47:30.886989 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.038061 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" event={"ID":"8f25ea0e-071a-4399-96f7-0c41367a232f","Type":"ContainerDied","Data":"6cff4f033810a27e6304bde3c50d6cb6d53186ffa1ecdae8b9529b78911d337d"} Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.038147 4706 scope.go:117] "RemoveContainer" containerID="ecb91bb37eeb4ff49424a5ed7f41da4468c7747d96e42b4f9f92207894d1af91" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.038369 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-4pxb7" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.045307 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-j2q44" event={"ID":"98651db7-0458-47ab-8621-7098bf5337ae","Type":"ContainerDied","Data":"4b460e45fc9bfe657f1c5945ab8f3de238db224c601a0210797fe7f3503bf975"} Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.045360 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b460e45fc9bfe657f1c5945ab8f3de238db224c601a0210797fe7f3503bf975" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.045472 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-j2q44" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.054234 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-sb\") pod \"8f25ea0e-071a-4399-96f7-0c41367a232f\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.054319 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-swift-storage-0\") pod \"8f25ea0e-071a-4399-96f7-0c41367a232f\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.054401 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-nb\") pod \"8f25ea0e-071a-4399-96f7-0c41367a232f\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.054433 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-openstack-edpm-ipam\") pod \"8f25ea0e-071a-4399-96f7-0c41367a232f\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.054561 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdwbk\" (UniqueName: \"kubernetes.io/projected/8f25ea0e-071a-4399-96f7-0c41367a232f-kube-api-access-sdwbk\") pod \"8f25ea0e-071a-4399-96f7-0c41367a232f\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.054681 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-svc\") pod \"8f25ea0e-071a-4399-96f7-0c41367a232f\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.054846 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-config\") pod \"8f25ea0e-071a-4399-96f7-0c41367a232f\" (UID: \"8f25ea0e-071a-4399-96f7-0c41367a232f\") " Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.067725 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f25ea0e-071a-4399-96f7-0c41367a232f-kube-api-access-sdwbk" (OuterVolumeSpecName: "kube-api-access-sdwbk") pod "8f25ea0e-071a-4399-96f7-0c41367a232f" (UID: "8f25ea0e-071a-4399-96f7-0c41367a232f"). InnerVolumeSpecName "kube-api-access-sdwbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.100660 4706 scope.go:117] "RemoveContainer" containerID="a758e87f1f4823ed673d33798d82f9b711d83da401be85ecf30a8f13a5323774" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.154162 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-config" (OuterVolumeSpecName: "config") pod "8f25ea0e-071a-4399-96f7-0c41367a232f" (UID: "8f25ea0e-071a-4399-96f7-0c41367a232f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.159497 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-config\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.159533 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdwbk\" (UniqueName: \"kubernetes.io/projected/8f25ea0e-071a-4399-96f7-0c41367a232f-kube-api-access-sdwbk\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.174755 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.175635 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" containerName="cloudkitty-proc" containerID="cri-o://0cb9a1ad5a1bb40669096ec41b53105becfc327ee89d251fa8c5f50db97e76b9" gracePeriod=30 Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.194190 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8f25ea0e-071a-4399-96f7-0c41367a232f" (UID: "8f25ea0e-071a-4399-96f7-0c41367a232f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.198272 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.199086 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api-log" containerID="cri-o://d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4" gracePeriod=30 Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.199320 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api" containerID="cri-o://5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849" gracePeriod=30 Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.202085 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "8f25ea0e-071a-4399-96f7-0c41367a232f" (UID: "8f25ea0e-071a-4399-96f7-0c41367a232f"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.239076 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8f25ea0e-071a-4399-96f7-0c41367a232f" (UID: "8f25ea0e-071a-4399-96f7-0c41367a232f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.257638 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8f25ea0e-071a-4399-96f7-0c41367a232f" (UID: "8f25ea0e-071a-4399-96f7-0c41367a232f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.261936 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8f25ea0e-071a-4399-96f7-0c41367a232f" (UID: "8f25ea0e-071a-4399-96f7-0c41367a232f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.263024 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.263086 4706 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.263139 4706 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.263157 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.263171 4706 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f25ea0e-071a-4399-96f7-0c41367a232f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.400283 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-4pxb7"] Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.409604 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-4pxb7"] Dec 08 19:47:31 crc kubenswrapper[4706]: I1208 19:47:31.627345 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f25ea0e-071a-4399-96f7-0c41367a232f" path="/var/lib/kubelet/pods/8f25ea0e-071a-4399-96f7-0c41367a232f/volumes" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.065272 4706 generic.go:334] "Generic (PLEG): container finished" podID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerID="d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4" exitCode=143 Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.065311 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"3863d194-54b8-4d2c-b605-4a6f19a51f49","Type":"ContainerDied","Data":"d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4"} Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.073005 4706 generic.go:334] "Generic (PLEG): container finished" podID="e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" containerID="0cb9a1ad5a1bb40669096ec41b53105becfc327ee89d251fa8c5f50db97e76b9" exitCode=0 Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.073067 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb","Type":"ContainerDied","Data":"0cb9a1ad5a1bb40669096ec41b53105becfc327ee89d251fa8c5f50db97e76b9"} Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.628714 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.708137 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-combined-ca-bundle\") pod \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.708355 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-scripts\") pod \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.708416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlhdc\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-kube-api-access-wlhdc\") pod \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.708636 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data\") pod \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.708869 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-certs\") pod \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.708937 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data-custom\") pod \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\" (UID: \"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.719945 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-scripts" (OuterVolumeSpecName: "scripts") pod "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" (UID: "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.721679 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-certs" (OuterVolumeSpecName: "certs") pod "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" (UID: "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.722980 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" (UID: "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.727180 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.729744 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-kube-api-access-wlhdc" (OuterVolumeSpecName: "kube-api-access-wlhdc") pod "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" (UID: "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb"). InnerVolumeSpecName "kube-api-access-wlhdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.819917 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-combined-ca-bundle\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.820039 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3863d194-54b8-4d2c-b605-4a6f19a51f49-logs\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.820209 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-public-tls-certs\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.820241 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-certs\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.820273 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-internal-tls-certs\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.820328 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-scripts\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.820351 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.820416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data-custom\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.820462 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5nz8\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-kube-api-access-d5nz8\") pod \"3863d194-54b8-4d2c-b605-4a6f19a51f49\" (UID: \"3863d194-54b8-4d2c-b605-4a6f19a51f49\") " Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.821104 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.821124 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.821136 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.821146 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlhdc\" (UniqueName: \"kubernetes.io/projected/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-kube-api-access-wlhdc\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.824489 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3863d194-54b8-4d2c-b605-4a6f19a51f49-logs" (OuterVolumeSpecName: "logs") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.827124 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" (UID: "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.830030 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data" (OuterVolumeSpecName: "config-data") pod "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" (UID: "e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.835949 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-scripts" (OuterVolumeSpecName: "scripts") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.836137 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-certs" (OuterVolumeSpecName: "certs") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.836576 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-kube-api-access-d5nz8" (OuterVolumeSpecName: "kube-api-access-d5nz8") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "kube-api-access-d5nz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.836625 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.905768 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.907681 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data" (OuterVolumeSpecName: "config-data") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925160 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925230 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925270 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3863d194-54b8-4d2c-b605-4a6f19a51f49-logs\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925301 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925314 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925343 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925375 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925385 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.925396 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5nz8\" (UniqueName: \"kubernetes.io/projected/3863d194-54b8-4d2c-b605-4a6f19a51f49-kube-api-access-d5nz8\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.946192 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:32 crc kubenswrapper[4706]: I1208 19:47:32.948797 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3863d194-54b8-4d2c-b605-4a6f19a51f49" (UID: "3863d194-54b8-4d2c-b605-4a6f19a51f49"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.027894 4706 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.027949 4706 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3863d194-54b8-4d2c-b605-4a6f19a51f49-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.090432 4706 generic.go:334] "Generic (PLEG): container finished" podID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerID="5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849" exitCode=0 Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.090540 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.090543 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"3863d194-54b8-4d2c-b605-4a6f19a51f49","Type":"ContainerDied","Data":"5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849"} Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.090674 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"3863d194-54b8-4d2c-b605-4a6f19a51f49","Type":"ContainerDied","Data":"5ed014b11d0d11b2a54ffafeac05bae67971e6fe47b2cb9fe987cb5f58470f2e"} Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.090721 4706 scope.go:117] "RemoveContainer" containerID="5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.093450 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb","Type":"ContainerDied","Data":"63b730bfd3410a162747fdaad00aa744d58eaa0c36222a7dea3b3fa467bd405a"} Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.093524 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.147563 4706 scope.go:117] "RemoveContainer" containerID="d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.157482 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.175451 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.189189 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.192116 4706 scope.go:117] "RemoveContainer" containerID="5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.193928 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849\": container with ID starting with 5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849 not found: ID does not exist" containerID="5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.193997 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849"} err="failed to get container status \"5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849\": rpc error: code = NotFound desc = could not find container \"5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849\": container with ID starting with 5c3061a0633e48cef4c59e09be339139c924b59e602851a51b05d4bf977e3849 not found: ID does not exist" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.194041 4706 scope.go:117] "RemoveContainer" containerID="d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.194677 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4\": container with ID starting with d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4 not found: ID does not exist" containerID="d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.194724 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4"} err="failed to get container status \"d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4\": rpc error: code = NotFound desc = could not find container \"d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4\": container with ID starting with d3909f95e35c16e239978711af09f8fb686215d630467cb2d7869b7ec85992b4 not found: ID does not exist" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.194746 4706 scope.go:117] "RemoveContainer" containerID="0cb9a1ad5a1bb40669096ec41b53105becfc327ee89d251fa8c5f50db97e76b9" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.215158 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.234970 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235514 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api-log" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235531 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api-log" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235550 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" containerName="cloudkitty-proc" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235558 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" containerName="cloudkitty-proc" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235567 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f25ea0e-071a-4399-96f7-0c41367a232f" containerName="dnsmasq-dns" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235577 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f25ea0e-071a-4399-96f7-0c41367a232f" containerName="dnsmasq-dns" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235617 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235625 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235640 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" containerName="extract-utilities" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235648 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" containerName="extract-utilities" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235665 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98651db7-0458-47ab-8621-7098bf5337ae" containerName="cloudkitty-storageinit" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235673 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="98651db7-0458-47ab-8621-7098bf5337ae" containerName="cloudkitty-storageinit" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235685 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" containerName="extract-content" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235692 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" containerName="extract-content" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235704 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f25ea0e-071a-4399-96f7-0c41367a232f" containerName="init" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235710 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f25ea0e-071a-4399-96f7-0c41367a232f" containerName="init" Dec 08 19:47:33 crc kubenswrapper[4706]: E1208 19:47:33.235729 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" containerName="registry-server" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235737 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" containerName="registry-server" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.235988 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" containerName="cloudkitty-proc" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.236011 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api-log" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.236019 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="98651db7-0458-47ab-8621-7098bf5337ae" containerName="cloudkitty-storageinit" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.236030 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f25ea0e-071a-4399-96f7-0c41367a232f" containerName="dnsmasq-dns" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.236054 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.236068 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd14ee90-6391-469b-b808-0a862f8683a5" containerName="registry-server" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.238377 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.243470 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.243678 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.243744 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-fzwvl" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.243778 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.243870 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.244324 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.244493 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.255472 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.273171 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.275687 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.279613 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.286208 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.344795 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.344893 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-scripts\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.345006 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.345065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjvqp\" (UniqueName: \"kubernetes.io/projected/b0316b44-87fb-44dd-890c-dace6fb0ab3e-kube-api-access-jjvqp\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.345133 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-config-data\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.345185 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.345222 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b0316b44-87fb-44dd-890c-dace6fb0ab3e-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.345285 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0316b44-87fb-44dd-890c-dace6fb0ab3e-logs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.345372 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.448505 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-certs\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.448567 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7w74\" (UniqueName: \"kubernetes.io/projected/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-kube-api-access-m7w74\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.448765 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.448880 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b0316b44-87fb-44dd-890c-dace6fb0ab3e-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.449101 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0316b44-87fb-44dd-890c-dace6fb0ab3e-logs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.449252 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-config-data\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.449420 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.449528 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.449651 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-scripts\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.449861 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.449903 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.449930 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.450011 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjvqp\" (UniqueName: \"kubernetes.io/projected/b0316b44-87fb-44dd-890c-dace6fb0ab3e-kube-api-access-jjvqp\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.450051 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-scripts\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.450433 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-config-data\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.451050 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0316b44-87fb-44dd-890c-dace6fb0ab3e-logs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.456073 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b0316b44-87fb-44dd-890c-dace6fb0ab3e-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.456845 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.457642 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.457720 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-scripts\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.459605 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-config-data\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.469443 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.470248 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0316b44-87fb-44dd-890c-dace6fb0ab3e-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.471174 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjvqp\" (UniqueName: \"kubernetes.io/projected/b0316b44-87fb-44dd-890c-dace6fb0ab3e-kube-api-access-jjvqp\") pod \"cloudkitty-api-0\" (UID: \"b0316b44-87fb-44dd-890c-dace6fb0ab3e\") " pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.554407 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.554486 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.554525 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-scripts\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.554574 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-certs\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.554596 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7w74\" (UniqueName: \"kubernetes.io/projected/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-kube-api-access-m7w74\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.554719 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-config-data\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.558803 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.560235 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-scripts\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.561019 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-config-data\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.564851 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-certs\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.565062 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.570858 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7w74\" (UniqueName: \"kubernetes.io/projected/ad4728d1-aa2e-4ab5-81dc-22f56a79e874-kube-api-access-m7w74\") pod \"cloudkitty-proc-0\" (UID: \"ad4728d1-aa2e-4ab5-81dc-22f56a79e874\") " pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.573225 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.607666 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.641183 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" path="/var/lib/kubelet/pods/3863d194-54b8-4d2c-b605-4a6f19a51f49/volumes" Dec 08 19:47:33 crc kubenswrapper[4706]: I1208 19:47:33.642048 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb" path="/var/lib/kubelet/pods/e460c85d-5d1f-4a1c-9cfc-0163f83a2ddb/volumes" Dec 08 19:47:34 crc kubenswrapper[4706]: I1208 19:47:34.119667 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 08 19:47:34 crc kubenswrapper[4706]: I1208 19:47:34.219523 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 08 19:47:35 crc kubenswrapper[4706]: I1208 19:47:35.154225 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"b0316b44-87fb-44dd-890c-dace6fb0ab3e","Type":"ContainerStarted","Data":"0378ec0687c7a7baa3ec19cabc4976b328c95a07999a3f482df2633b26fdb658"} Dec 08 19:47:35 crc kubenswrapper[4706]: I1208 19:47:35.155111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"b0316b44-87fb-44dd-890c-dace6fb0ab3e","Type":"ContainerStarted","Data":"1da5ef54b69eed06b99c515fcc1fde564e028a55df0bf57e58064db7e364d1c2"} Dec 08 19:47:35 crc kubenswrapper[4706]: I1208 19:47:35.155125 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"b0316b44-87fb-44dd-890c-dace6fb0ab3e","Type":"ContainerStarted","Data":"6fc365ef49a80b14d43a25de828d2ce658d472a7024c5650826d16cc184fa408"} Dec 08 19:47:35 crc kubenswrapper[4706]: I1208 19:47:35.157436 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 08 19:47:35 crc kubenswrapper[4706]: I1208 19:47:35.164578 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"ad4728d1-aa2e-4ab5-81dc-22f56a79e874","Type":"ContainerStarted","Data":"a0c5531fd739ce1e2e024c38243dc99065454e463e74b47a3b9a707a07a218fd"} Dec 08 19:47:35 crc kubenswrapper[4706]: I1208 19:47:35.218891 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.2188684260000002 podStartE2EDuration="2.218868426s" podCreationTimestamp="2025-12-08 19:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:47:35.214670186 +0000 UTC m=+1537.856871189" watchObservedRunningTime="2025-12-08 19:47:35.218868426 +0000 UTC m=+1537.861069429" Dec 08 19:47:36 crc kubenswrapper[4706]: I1208 19:47:36.184540 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"ad4728d1-aa2e-4ab5-81dc-22f56a79e874","Type":"ContainerStarted","Data":"cf3aa2f9f944da5a112b14e8ccae4445ebb317b286c1fc3c8d45f3c20edde8b5"} Dec 08 19:47:36 crc kubenswrapper[4706]: I1208 19:47:36.224677 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=1.809537865 podStartE2EDuration="3.224647711s" podCreationTimestamp="2025-12-08 19:47:33 +0000 UTC" firstStartedPulling="2025-12-08 19:47:34.221934973 +0000 UTC m=+1536.864135976" lastFinishedPulling="2025-12-08 19:47:35.637044819 +0000 UTC m=+1538.279245822" observedRunningTime="2025-12-08 19:47:36.208181603 +0000 UTC m=+1538.850382616" watchObservedRunningTime="2025-12-08 19:47:36.224647711 +0000 UTC m=+1538.866848714" Dec 08 19:47:37 crc kubenswrapper[4706]: I1208 19:47:37.623106 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-api-0" podUID="3863d194-54b8-4d2c-b605-4a6f19a51f49" containerName="cloudkitty-api" probeResult="failure" output="Get \"https://10.217.0.186:8889/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.721681 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6"] Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.724181 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.727713 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.727827 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.727948 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.729720 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.749754 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6"] Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.807012 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.807123 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.807193 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.807253 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sshtg\" (UniqueName: \"kubernetes.io/projected/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-kube-api-access-sshtg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.910647 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.910735 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.910790 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.910847 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sshtg\" (UniqueName: \"kubernetes.io/projected/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-kube-api-access-sshtg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.918495 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.921116 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.923556 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:38 crc kubenswrapper[4706]: I1208 19:47:38.939742 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sshtg\" (UniqueName: \"kubernetes.io/projected/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-kube-api-access-sshtg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:39 crc kubenswrapper[4706]: I1208 19:47:39.047250 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:47:39 crc kubenswrapper[4706]: I1208 19:47:39.682821 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6"] Dec 08 19:47:39 crc kubenswrapper[4706]: I1208 19:47:39.688000 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 19:47:40 crc kubenswrapper[4706]: I1208 19:47:40.252297 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" event={"ID":"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75","Type":"ContainerStarted","Data":"22174f1100eee236202b9cb62f31e6a7e2572a506f28ac41c933f98671e0cc24"} Dec 08 19:47:45 crc kubenswrapper[4706]: I1208 19:47:45.349773 4706 generic.go:334] "Generic (PLEG): container finished" podID="ecc9ad61-4d9c-4d26-a240-398cfc8c3165" containerID="1bbcfb38eef75278ddd8dcd66ef17d433522e1fdce4e1162896c39cbc0ac2fdb" exitCode=0 Dec 08 19:47:45 crc kubenswrapper[4706]: I1208 19:47:45.349927 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecc9ad61-4d9c-4d26-a240-398cfc8c3165","Type":"ContainerDied","Data":"1bbcfb38eef75278ddd8dcd66ef17d433522e1fdce4e1162896c39cbc0ac2fdb"} Dec 08 19:47:45 crc kubenswrapper[4706]: I1208 19:47:45.353418 4706 generic.go:334] "Generic (PLEG): container finished" podID="faed709c-224b-4c30-9af3-d64d34525599" containerID="f8ea3e55f55046de20a740b5a047ac905eeae4750bb017877973debfa7058da0" exitCode=0 Dec 08 19:47:45 crc kubenswrapper[4706]: I1208 19:47:45.353474 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"faed709c-224b-4c30-9af3-d64d34525599","Type":"ContainerDied","Data":"f8ea3e55f55046de20a740b5a047ac905eeae4750bb017877973debfa7058da0"} Dec 08 19:47:48 crc kubenswrapper[4706]: I1208 19:47:48.823241 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 08 19:47:50 crc kubenswrapper[4706]: I1208 19:47:50.432009 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"faed709c-224b-4c30-9af3-d64d34525599","Type":"ContainerStarted","Data":"5c1ec71930dbe16a8db293b4b93ad3073dbdcdf9fd7d88262080b06ab16318d9"} Dec 08 19:47:50 crc kubenswrapper[4706]: I1208 19:47:50.435227 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 08 19:47:50 crc kubenswrapper[4706]: I1208 19:47:50.440108 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecc9ad61-4d9c-4d26-a240-398cfc8c3165","Type":"ContainerStarted","Data":"4cbbadf83ea6039a12309b7bc66c75f520e69e85557f545465d2419ebe10bb51"} Dec 08 19:47:50 crc kubenswrapper[4706]: I1208 19:47:50.440446 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:47:50 crc kubenswrapper[4706]: I1208 19:47:50.444658 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" event={"ID":"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75","Type":"ContainerStarted","Data":"990b5fc11f8d7c2f1d0842bd0db98f5e2b469dcdffda7d963a0cd0f83303d218"} Dec 08 19:47:50 crc kubenswrapper[4706]: I1208 19:47:50.470782 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=48.470760635 podStartE2EDuration="48.470760635s" podCreationTimestamp="2025-12-08 19:47:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:47:50.469849439 +0000 UTC m=+1553.112050442" watchObservedRunningTime="2025-12-08 19:47:50.470760635 +0000 UTC m=+1553.112961638" Dec 08 19:47:50 crc kubenswrapper[4706]: I1208 19:47:50.493471 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" podStartSLOduration=2.734365033 podStartE2EDuration="12.493450151s" podCreationTimestamp="2025-12-08 19:47:38 +0000 UTC" firstStartedPulling="2025-12-08 19:47:39.687696534 +0000 UTC m=+1542.329897537" lastFinishedPulling="2025-12-08 19:47:49.446781642 +0000 UTC m=+1552.088982655" observedRunningTime="2025-12-08 19:47:50.492864184 +0000 UTC m=+1553.135065197" watchObservedRunningTime="2025-12-08 19:47:50.493450151 +0000 UTC m=+1553.135651154" Dec 08 19:47:50 crc kubenswrapper[4706]: I1208 19:47:50.532860 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=45.53282325 podStartE2EDuration="45.53282325s" podCreationTimestamp="2025-12-08 19:47:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 19:47:50.516686781 +0000 UTC m=+1553.158887784" watchObservedRunningTime="2025-12-08 19:47:50.53282325 +0000 UTC m=+1553.175024253" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.534413 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8t55g"] Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.537662 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.570365 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t55g"] Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.598849 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-utilities\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.598950 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v2cw\" (UniqueName: \"kubernetes.io/projected/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-kube-api-access-7v2cw\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.599000 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-catalog-content\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.702279 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-utilities\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.702972 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-utilities\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.703432 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v2cw\" (UniqueName: \"kubernetes.io/projected/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-kube-api-access-7v2cw\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.704338 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-catalog-content\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.704552 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-catalog-content\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.726093 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v2cw\" (UniqueName: \"kubernetes.io/projected/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-kube-api-access-7v2cw\") pod \"redhat-marketplace-8t55g\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:52 crc kubenswrapper[4706]: I1208 19:47:52.861025 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:47:53 crc kubenswrapper[4706]: I1208 19:47:53.367995 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t55g"] Dec 08 19:47:53 crc kubenswrapper[4706]: I1208 19:47:53.484407 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t55g" event={"ID":"9bc99bda-df61-4bb2-8eed-9e54cf51ce84","Type":"ContainerStarted","Data":"29fb15f9711d94c387944ab9f9be3ce6d560d620de4b884396e9dae8b936bcc8"} Dec 08 19:47:54 crc kubenswrapper[4706]: I1208 19:47:54.498074 4706 generic.go:334] "Generic (PLEG): container finished" podID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerID="7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22" exitCode=0 Dec 08 19:47:54 crc kubenswrapper[4706]: I1208 19:47:54.498188 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t55g" event={"ID":"9bc99bda-df61-4bb2-8eed-9e54cf51ce84","Type":"ContainerDied","Data":"7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22"} Dec 08 19:47:58 crc kubenswrapper[4706]: I1208 19:47:58.547299 4706 generic.go:334] "Generic (PLEG): container finished" podID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerID="2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861" exitCode=0 Dec 08 19:47:58 crc kubenswrapper[4706]: I1208 19:47:58.547423 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t55g" event={"ID":"9bc99bda-df61-4bb2-8eed-9e54cf51ce84","Type":"ContainerDied","Data":"2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861"} Dec 08 19:48:00 crc kubenswrapper[4706]: I1208 19:48:00.576460 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t55g" event={"ID":"9bc99bda-df61-4bb2-8eed-9e54cf51ce84","Type":"ContainerStarted","Data":"638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93"} Dec 08 19:48:00 crc kubenswrapper[4706]: I1208 19:48:00.613160 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8t55g" podStartSLOduration=3.15614695 podStartE2EDuration="8.613132042s" podCreationTimestamp="2025-12-08 19:47:52 +0000 UTC" firstStartedPulling="2025-12-08 19:47:54.502381639 +0000 UTC m=+1557.144582642" lastFinishedPulling="2025-12-08 19:47:59.959366731 +0000 UTC m=+1562.601567734" observedRunningTime="2025-12-08 19:48:00.600311366 +0000 UTC m=+1563.242512369" watchObservedRunningTime="2025-12-08 19:48:00.613132042 +0000 UTC m=+1563.255333045" Dec 08 19:48:01 crc kubenswrapper[4706]: I1208 19:48:01.591887 4706 generic.go:334] "Generic (PLEG): container finished" podID="ad5354c5-7ee5-4e8a-8e20-84ac36e97e75" containerID="990b5fc11f8d7c2f1d0842bd0db98f5e2b469dcdffda7d963a0cd0f83303d218" exitCode=0 Dec 08 19:48:01 crc kubenswrapper[4706]: I1208 19:48:01.593893 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" event={"ID":"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75","Type":"ContainerDied","Data":"990b5fc11f8d7c2f1d0842bd0db98f5e2b469dcdffda7d963a0cd0f83303d218"} Dec 08 19:48:02 crc kubenswrapper[4706]: I1208 19:48:02.861823 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:48:02 crc kubenswrapper[4706]: I1208 19:48:02.862506 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:48:02 crc kubenswrapper[4706]: I1208 19:48:02.921195 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:48:02 crc kubenswrapper[4706]: I1208 19:48:02.956038 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.192626 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.303197 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-inventory\") pod \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.303316 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-repo-setup-combined-ca-bundle\") pod \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.303532 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-ssh-key\") pod \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.303582 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sshtg\" (UniqueName: \"kubernetes.io/projected/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-kube-api-access-sshtg\") pod \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\" (UID: \"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75\") " Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.330532 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-kube-api-access-sshtg" (OuterVolumeSpecName: "kube-api-access-sshtg") pod "ad5354c5-7ee5-4e8a-8e20-84ac36e97e75" (UID: "ad5354c5-7ee5-4e8a-8e20-84ac36e97e75"). InnerVolumeSpecName "kube-api-access-sshtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.363458 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "ad5354c5-7ee5-4e8a-8e20-84ac36e97e75" (UID: "ad5354c5-7ee5-4e8a-8e20-84ac36e97e75"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.407931 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sshtg\" (UniqueName: \"kubernetes.io/projected/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-kube-api-access-sshtg\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.408403 4706 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.419471 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ad5354c5-7ee5-4e8a-8e20-84ac36e97e75" (UID: "ad5354c5-7ee5-4e8a-8e20-84ac36e97e75"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.497445 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-inventory" (OuterVolumeSpecName: "inventory") pod "ad5354c5-7ee5-4e8a-8e20-84ac36e97e75" (UID: "ad5354c5-7ee5-4e8a-8e20-84ac36e97e75"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.510895 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.510940 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad5354c5-7ee5-4e8a-8e20-84ac36e97e75-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.633569 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" event={"ID":"ad5354c5-7ee5-4e8a-8e20-84ac36e97e75","Type":"ContainerDied","Data":"22174f1100eee236202b9cb62f31e6a7e2572a506f28ac41c933f98671e0cc24"} Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.633654 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22174f1100eee236202b9cb62f31e6a7e2572a506f28ac41c933f98671e0cc24" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.633603 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.734109 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v"] Dec 08 19:48:03 crc kubenswrapper[4706]: E1208 19:48:03.734721 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad5354c5-7ee5-4e8a-8e20-84ac36e97e75" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.734743 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad5354c5-7ee5-4e8a-8e20-84ac36e97e75" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.735195 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad5354c5-7ee5-4e8a-8e20-84ac36e97e75" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.736787 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.739698 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.739882 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.740771 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.740969 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.744675 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v"] Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.819159 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.819226 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsvfm\" (UniqueName: \"kubernetes.io/projected/074561a7-7190-4458-a9aa-1dac832175fb-kube-api-access-jsvfm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.820301 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.922576 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.922673 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsvfm\" (UniqueName: \"kubernetes.io/projected/074561a7-7190-4458-a9aa-1dac832175fb-kube-api-access-jsvfm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.922935 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.927958 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.928089 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:03 crc kubenswrapper[4706]: I1208 19:48:03.940371 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsvfm\" (UniqueName: \"kubernetes.io/projected/074561a7-7190-4458-a9aa-1dac832175fb-kube-api-access-jsvfm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hvs9v\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:04 crc kubenswrapper[4706]: I1208 19:48:04.076725 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:04 crc kubenswrapper[4706]: I1208 19:48:04.723798 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v"] Dec 08 19:48:05 crc kubenswrapper[4706]: I1208 19:48:05.664784 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" event={"ID":"074561a7-7190-4458-a9aa-1dac832175fb","Type":"ContainerStarted","Data":"df2b30e173e0053185ee5b46003822f505e14761521ce61081b15858c544dd63"} Dec 08 19:48:05 crc kubenswrapper[4706]: I1208 19:48:05.665219 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" event={"ID":"074561a7-7190-4458-a9aa-1dac832175fb","Type":"ContainerStarted","Data":"4dd0a6106208eeaedcb5fe6c9a4a29e876884d264b279a30ae156425b1aef884"} Dec 08 19:48:05 crc kubenswrapper[4706]: I1208 19:48:05.686536 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" podStartSLOduration=2.250040526 podStartE2EDuration="2.686509435s" podCreationTimestamp="2025-12-08 19:48:03 +0000 UTC" firstStartedPulling="2025-12-08 19:48:04.735170013 +0000 UTC m=+1567.377371016" lastFinishedPulling="2025-12-08 19:48:05.171638912 +0000 UTC m=+1567.813839925" observedRunningTime="2025-12-08 19:48:05.684769846 +0000 UTC m=+1568.326970849" watchObservedRunningTime="2025-12-08 19:48:05.686509435 +0000 UTC m=+1568.328710468" Dec 08 19:48:06 crc kubenswrapper[4706]: I1208 19:48:06.297634 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 08 19:48:08 crc kubenswrapper[4706]: I1208 19:48:08.738932 4706 generic.go:334] "Generic (PLEG): container finished" podID="074561a7-7190-4458-a9aa-1dac832175fb" containerID="df2b30e173e0053185ee5b46003822f505e14761521ce61081b15858c544dd63" exitCode=0 Dec 08 19:48:08 crc kubenswrapper[4706]: I1208 19:48:08.739630 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" event={"ID":"074561a7-7190-4458-a9aa-1dac832175fb","Type":"ContainerDied","Data":"df2b30e173e0053185ee5b46003822f505e14761521ce61081b15858c544dd63"} Dec 08 19:48:09 crc kubenswrapper[4706]: I1208 19:48:09.564439 4706 scope.go:117] "RemoveContainer" containerID="bc2a3c7a3ebbe36ae3f29c3fdd4bc71d23a486cf3882c63ecb8cfb1c56b6ab4f" Dec 08 19:48:09 crc kubenswrapper[4706]: I1208 19:48:09.614692 4706 scope.go:117] "RemoveContainer" containerID="927f5c41ed95fe2418cd7edb1c7e27514c71e951fd9dc76c0b08c1bd76c67e93" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.388945 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.517916 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsvfm\" (UniqueName: \"kubernetes.io/projected/074561a7-7190-4458-a9aa-1dac832175fb-kube-api-access-jsvfm\") pod \"074561a7-7190-4458-a9aa-1dac832175fb\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.518435 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-inventory\") pod \"074561a7-7190-4458-a9aa-1dac832175fb\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.518607 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-ssh-key\") pod \"074561a7-7190-4458-a9aa-1dac832175fb\" (UID: \"074561a7-7190-4458-a9aa-1dac832175fb\") " Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.541408 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/074561a7-7190-4458-a9aa-1dac832175fb-kube-api-access-jsvfm" (OuterVolumeSpecName: "kube-api-access-jsvfm") pod "074561a7-7190-4458-a9aa-1dac832175fb" (UID: "074561a7-7190-4458-a9aa-1dac832175fb"). InnerVolumeSpecName "kube-api-access-jsvfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.559043 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-inventory" (OuterVolumeSpecName: "inventory") pod "074561a7-7190-4458-a9aa-1dac832175fb" (UID: "074561a7-7190-4458-a9aa-1dac832175fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.565236 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "074561a7-7190-4458-a9aa-1dac832175fb" (UID: "074561a7-7190-4458-a9aa-1dac832175fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.623868 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsvfm\" (UniqueName: \"kubernetes.io/projected/074561a7-7190-4458-a9aa-1dac832175fb-kube-api-access-jsvfm\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.623912 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.623926 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/074561a7-7190-4458-a9aa-1dac832175fb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.637982 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.780913 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" event={"ID":"074561a7-7190-4458-a9aa-1dac832175fb","Type":"ContainerDied","Data":"4dd0a6106208eeaedcb5fe6c9a4a29e876884d264b279a30ae156425b1aef884"} Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.781368 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dd0a6106208eeaedcb5fe6c9a4a29e876884d264b279a30ae156425b1aef884" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.781436 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hvs9v" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.859824 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl"] Dec 08 19:48:10 crc kubenswrapper[4706]: E1208 19:48:10.860485 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074561a7-7190-4458-a9aa-1dac832175fb" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.860509 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="074561a7-7190-4458-a9aa-1dac832175fb" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.860783 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="074561a7-7190-4458-a9aa-1dac832175fb" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.861783 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.865876 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.866028 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.866043 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.866028 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.873237 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl"] Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.932690 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk6gd\" (UniqueName: \"kubernetes.io/projected/91cd912d-3492-4709-99e9-94fa496bde14-kube-api-access-lk6gd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.933132 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.933755 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:10 crc kubenswrapper[4706]: I1208 19:48:10.933936 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.036710 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.036788 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.036858 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk6gd\" (UniqueName: \"kubernetes.io/projected/91cd912d-3492-4709-99e9-94fa496bde14-kube-api-access-lk6gd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.036940 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.043512 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.044746 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.051420 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.060068 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk6gd\" (UniqueName: \"kubernetes.io/projected/91cd912d-3492-4709-99e9-94fa496bde14-kube-api-access-lk6gd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-267vl\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.198815 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:48:11 crc kubenswrapper[4706]: I1208 19:48:11.831322 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl"] Dec 08 19:48:11 crc kubenswrapper[4706]: W1208 19:48:11.844659 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91cd912d_3492_4709_99e9_94fa496bde14.slice/crio-9c227434af50512f5071448c3f00a2b28b56966f7f94e154f6286088ba52ea82 WatchSource:0}: Error finding container 9c227434af50512f5071448c3f00a2b28b56966f7f94e154f6286088ba52ea82: Status 404 returned error can't find the container with id 9c227434af50512f5071448c3f00a2b28b56966f7f94e154f6286088ba52ea82 Dec 08 19:48:12 crc kubenswrapper[4706]: I1208 19:48:12.809167 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" event={"ID":"91cd912d-3492-4709-99e9-94fa496bde14","Type":"ContainerStarted","Data":"63bfb742e3057571684fdd52cbe981af33b67db6bbdf7e574dfab7a8bddb4206"} Dec 08 19:48:12 crc kubenswrapper[4706]: I1208 19:48:12.809973 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" event={"ID":"91cd912d-3492-4709-99e9-94fa496bde14","Type":"ContainerStarted","Data":"9c227434af50512f5071448c3f00a2b28b56966f7f94e154f6286088ba52ea82"} Dec 08 19:48:12 crc kubenswrapper[4706]: I1208 19:48:12.834638 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" podStartSLOduration=2.367292446 podStartE2EDuration="2.834598134s" podCreationTimestamp="2025-12-08 19:48:10 +0000 UTC" firstStartedPulling="2025-12-08 19:48:11.850140778 +0000 UTC m=+1574.492341781" lastFinishedPulling="2025-12-08 19:48:12.317446466 +0000 UTC m=+1574.959647469" observedRunningTime="2025-12-08 19:48:12.827011158 +0000 UTC m=+1575.469212161" watchObservedRunningTime="2025-12-08 19:48:12.834598134 +0000 UTC m=+1575.476799137" Dec 08 19:48:12 crc kubenswrapper[4706]: I1208 19:48:12.922304 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:48:12 crc kubenswrapper[4706]: I1208 19:48:12.981175 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t55g"] Dec 08 19:48:13 crc kubenswrapper[4706]: I1208 19:48:13.819901 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8t55g" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerName="registry-server" containerID="cri-o://638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93" gracePeriod=2 Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.477931 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.631558 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-catalog-content\") pod \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.632306 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-utilities\") pod \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.632392 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v2cw\" (UniqueName: \"kubernetes.io/projected/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-kube-api-access-7v2cw\") pod \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\" (UID: \"9bc99bda-df61-4bb2-8eed-9e54cf51ce84\") " Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.632964 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-utilities" (OuterVolumeSpecName: "utilities") pod "9bc99bda-df61-4bb2-8eed-9e54cf51ce84" (UID: "9bc99bda-df61-4bb2-8eed-9e54cf51ce84"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.634108 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.642067 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-kube-api-access-7v2cw" (OuterVolumeSpecName: "kube-api-access-7v2cw") pod "9bc99bda-df61-4bb2-8eed-9e54cf51ce84" (UID: "9bc99bda-df61-4bb2-8eed-9e54cf51ce84"). InnerVolumeSpecName "kube-api-access-7v2cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.651799 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bc99bda-df61-4bb2-8eed-9e54cf51ce84" (UID: "9bc99bda-df61-4bb2-8eed-9e54cf51ce84"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.736571 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.736612 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v2cw\" (UniqueName: \"kubernetes.io/projected/9bc99bda-df61-4bb2-8eed-9e54cf51ce84-kube-api-access-7v2cw\") on node \"crc\" DevicePath \"\"" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.843387 4706 generic.go:334] "Generic (PLEG): container finished" podID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerID="638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93" exitCode=0 Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.843442 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t55g" event={"ID":"9bc99bda-df61-4bb2-8eed-9e54cf51ce84","Type":"ContainerDied","Data":"638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93"} Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.843481 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t55g" event={"ID":"9bc99bda-df61-4bb2-8eed-9e54cf51ce84","Type":"ContainerDied","Data":"29fb15f9711d94c387944ab9f9be3ce6d560d620de4b884396e9dae8b936bcc8"} Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.843509 4706 scope.go:117] "RemoveContainer" containerID="638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.843734 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8t55g" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.891081 4706 scope.go:117] "RemoveContainer" containerID="2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.904979 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t55g"] Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.927306 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t55g"] Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.935535 4706 scope.go:117] "RemoveContainer" containerID="7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.981353 4706 scope.go:117] "RemoveContainer" containerID="638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93" Dec 08 19:48:14 crc kubenswrapper[4706]: E1208 19:48:14.982311 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93\": container with ID starting with 638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93 not found: ID does not exist" containerID="638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.982370 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93"} err="failed to get container status \"638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93\": rpc error: code = NotFound desc = could not find container \"638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93\": container with ID starting with 638e5b117c5866d3b864ef790b1edf2eb3ae639fa845ab7167c41e9cee816d93 not found: ID does not exist" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.982405 4706 scope.go:117] "RemoveContainer" containerID="2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861" Dec 08 19:48:14 crc kubenswrapper[4706]: E1208 19:48:14.984051 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861\": container with ID starting with 2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861 not found: ID does not exist" containerID="2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.984103 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861"} err="failed to get container status \"2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861\": rpc error: code = NotFound desc = could not find container \"2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861\": container with ID starting with 2b30075d3d7f578de004aa40e1b1331ba02bca472a3e577454e40ddca7400861 not found: ID does not exist" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.984147 4706 scope.go:117] "RemoveContainer" containerID="7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22" Dec 08 19:48:14 crc kubenswrapper[4706]: E1208 19:48:14.986028 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22\": container with ID starting with 7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22 not found: ID does not exist" containerID="7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22" Dec 08 19:48:14 crc kubenswrapper[4706]: I1208 19:48:14.986061 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22"} err="failed to get container status \"7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22\": rpc error: code = NotFound desc = could not find container \"7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22\": container with ID starting with 7965f40271d8ae457e4096d91e5337481545c6eae6ef8a420c4d30aad17c0d22 not found: ID does not exist" Dec 08 19:48:15 crc kubenswrapper[4706]: I1208 19:48:15.623510 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" path="/var/lib/kubelet/pods/9bc99bda-df61-4bb2-8eed-9e54cf51ce84/volumes" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.461969 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-29bp4"] Dec 08 19:48:49 crc kubenswrapper[4706]: E1208 19:48:49.463461 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerName="extract-content" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.463482 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerName="extract-content" Dec 08 19:48:49 crc kubenswrapper[4706]: E1208 19:48:49.463527 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerName="registry-server" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.463535 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerName="registry-server" Dec 08 19:48:49 crc kubenswrapper[4706]: E1208 19:48:49.463557 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerName="extract-utilities" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.463587 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerName="extract-utilities" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.463833 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc99bda-df61-4bb2-8eed-9e54cf51ce84" containerName="registry-server" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.466306 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.480015 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-29bp4"] Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.553387 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-utilities\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.553652 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-catalog-content\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.554434 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpf75\" (UniqueName: \"kubernetes.io/projected/2c91542a-001c-48aa-9caf-94d2a35af371-kube-api-access-tpf75\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.656794 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpf75\" (UniqueName: \"kubernetes.io/projected/2c91542a-001c-48aa-9caf-94d2a35af371-kube-api-access-tpf75\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.656901 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-utilities\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.657001 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-catalog-content\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.657488 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-utilities\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.657893 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-catalog-content\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.680329 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpf75\" (UniqueName: \"kubernetes.io/projected/2c91542a-001c-48aa-9caf-94d2a35af371-kube-api-access-tpf75\") pod \"community-operators-29bp4\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:49 crc kubenswrapper[4706]: I1208 19:48:49.795582 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:50 crc kubenswrapper[4706]: I1208 19:48:50.317985 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-29bp4"] Dec 08 19:48:51 crc kubenswrapper[4706]: I1208 19:48:51.297990 4706 generic.go:334] "Generic (PLEG): container finished" podID="2c91542a-001c-48aa-9caf-94d2a35af371" containerID="35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028" exitCode=0 Dec 08 19:48:51 crc kubenswrapper[4706]: I1208 19:48:51.298219 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29bp4" event={"ID":"2c91542a-001c-48aa-9caf-94d2a35af371","Type":"ContainerDied","Data":"35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028"} Dec 08 19:48:51 crc kubenswrapper[4706]: I1208 19:48:51.298456 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29bp4" event={"ID":"2c91542a-001c-48aa-9caf-94d2a35af371","Type":"ContainerStarted","Data":"5e0d1486b7183d5bacb8d0fe4b2dc588019b58282a851e6f5e6ec8e2b1213e4c"} Dec 08 19:48:52 crc kubenswrapper[4706]: I1208 19:48:52.312958 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29bp4" event={"ID":"2c91542a-001c-48aa-9caf-94d2a35af371","Type":"ContainerStarted","Data":"da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d"} Dec 08 19:48:53 crc kubenswrapper[4706]: I1208 19:48:53.327946 4706 generic.go:334] "Generic (PLEG): container finished" podID="2c91542a-001c-48aa-9caf-94d2a35af371" containerID="da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d" exitCode=0 Dec 08 19:48:53 crc kubenswrapper[4706]: I1208 19:48:53.328172 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29bp4" event={"ID":"2c91542a-001c-48aa-9caf-94d2a35af371","Type":"ContainerDied","Data":"da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d"} Dec 08 19:48:54 crc kubenswrapper[4706]: I1208 19:48:54.369910 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29bp4" event={"ID":"2c91542a-001c-48aa-9caf-94d2a35af371","Type":"ContainerStarted","Data":"97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6"} Dec 08 19:48:54 crc kubenswrapper[4706]: I1208 19:48:54.403690 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-29bp4" podStartSLOduration=2.937949055 podStartE2EDuration="5.403654194s" podCreationTimestamp="2025-12-08 19:48:49 +0000 UTC" firstStartedPulling="2025-12-08 19:48:51.301384145 +0000 UTC m=+1613.943585148" lastFinishedPulling="2025-12-08 19:48:53.767089284 +0000 UTC m=+1616.409290287" observedRunningTime="2025-12-08 19:48:54.393559997 +0000 UTC m=+1617.035761010" watchObservedRunningTime="2025-12-08 19:48:54.403654194 +0000 UTC m=+1617.045855197" Dec 08 19:48:59 crc kubenswrapper[4706]: I1208 19:48:59.795978 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:59 crc kubenswrapper[4706]: I1208 19:48:59.796724 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:48:59 crc kubenswrapper[4706]: I1208 19:48:59.861008 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:49:00 crc kubenswrapper[4706]: I1208 19:49:00.490539 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:49:00 crc kubenswrapper[4706]: I1208 19:49:00.552773 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-29bp4"] Dec 08 19:49:02 crc kubenswrapper[4706]: I1208 19:49:02.475819 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-29bp4" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" containerName="registry-server" containerID="cri-o://97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6" gracePeriod=2 Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.209625 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.326387 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpf75\" (UniqueName: \"kubernetes.io/projected/2c91542a-001c-48aa-9caf-94d2a35af371-kube-api-access-tpf75\") pod \"2c91542a-001c-48aa-9caf-94d2a35af371\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.326660 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-catalog-content\") pod \"2c91542a-001c-48aa-9caf-94d2a35af371\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.326896 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-utilities\") pod \"2c91542a-001c-48aa-9caf-94d2a35af371\" (UID: \"2c91542a-001c-48aa-9caf-94d2a35af371\") " Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.327858 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-utilities" (OuterVolumeSpecName: "utilities") pod "2c91542a-001c-48aa-9caf-94d2a35af371" (UID: "2c91542a-001c-48aa-9caf-94d2a35af371"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.335409 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c91542a-001c-48aa-9caf-94d2a35af371-kube-api-access-tpf75" (OuterVolumeSpecName: "kube-api-access-tpf75") pod "2c91542a-001c-48aa-9caf-94d2a35af371" (UID: "2c91542a-001c-48aa-9caf-94d2a35af371"). InnerVolumeSpecName "kube-api-access-tpf75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.379924 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c91542a-001c-48aa-9caf-94d2a35af371" (UID: "2c91542a-001c-48aa-9caf-94d2a35af371"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.429957 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.429996 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c91542a-001c-48aa-9caf-94d2a35af371-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.430007 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpf75\" (UniqueName: \"kubernetes.io/projected/2c91542a-001c-48aa-9caf-94d2a35af371-kube-api-access-tpf75\") on node \"crc\" DevicePath \"\"" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.489687 4706 generic.go:334] "Generic (PLEG): container finished" podID="2c91542a-001c-48aa-9caf-94d2a35af371" containerID="97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6" exitCode=0 Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.489750 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29bp4" event={"ID":"2c91542a-001c-48aa-9caf-94d2a35af371","Type":"ContainerDied","Data":"97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6"} Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.489776 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29bp4" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.489803 4706 scope.go:117] "RemoveContainer" containerID="97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.489789 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29bp4" event={"ID":"2c91542a-001c-48aa-9caf-94d2a35af371","Type":"ContainerDied","Data":"5e0d1486b7183d5bacb8d0fe4b2dc588019b58282a851e6f5e6ec8e2b1213e4c"} Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.522679 4706 scope.go:117] "RemoveContainer" containerID="da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.533575 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-29bp4"] Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.543332 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-29bp4"] Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.565176 4706 scope.go:117] "RemoveContainer" containerID="35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.621395 4706 scope.go:117] "RemoveContainer" containerID="97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6" Dec 08 19:49:03 crc kubenswrapper[4706]: E1208 19:49:03.624967 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6\": container with ID starting with 97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6 not found: ID does not exist" containerID="97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.625040 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6"} err="failed to get container status \"97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6\": rpc error: code = NotFound desc = could not find container \"97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6\": container with ID starting with 97503febdb2be3c5cfabe24dff3c9c716b3c8f7c34f49faa131794e48b6dc6a6 not found: ID does not exist" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.625090 4706 scope.go:117] "RemoveContainer" containerID="da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d" Dec 08 19:49:03 crc kubenswrapper[4706]: E1208 19:49:03.625714 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d\": container with ID starting with da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d not found: ID does not exist" containerID="da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.625791 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d"} err="failed to get container status \"da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d\": rpc error: code = NotFound desc = could not find container \"da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d\": container with ID starting with da36c45b95b4ac22c162b96c26a3bd31905f3509d71a97527d3e684e27daf85d not found: ID does not exist" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.625838 4706 scope.go:117] "RemoveContainer" containerID="35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028" Dec 08 19:49:03 crc kubenswrapper[4706]: E1208 19:49:03.626199 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028\": container with ID starting with 35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028 not found: ID does not exist" containerID="35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.626231 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028"} err="failed to get container status \"35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028\": rpc error: code = NotFound desc = could not find container \"35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028\": container with ID starting with 35a5dc722890dd25971142fcef64635f42926b2362a7f9e7225d67169b5b6028 not found: ID does not exist" Dec 08 19:49:03 crc kubenswrapper[4706]: I1208 19:49:03.628704 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" path="/var/lib/kubelet/pods/2c91542a-001c-48aa-9caf-94d2a35af371/volumes" Dec 08 19:49:05 crc kubenswrapper[4706]: I1208 19:49:05.836088 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:49:05 crc kubenswrapper[4706]: I1208 19:49:05.837082 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:49:09 crc kubenswrapper[4706]: I1208 19:49:09.870577 4706 scope.go:117] "RemoveContainer" containerID="d6b9699ca260769714dd92bd847e10ad0349af7a1c4db2688c6afe4a9dd0a788" Dec 08 19:49:09 crc kubenswrapper[4706]: I1208 19:49:09.924367 4706 scope.go:117] "RemoveContainer" containerID="891fc2499228fc87fb1799688a10e66c513f244ea0bbbe204a8c839d63f4e8ce" Dec 08 19:49:35 crc kubenswrapper[4706]: I1208 19:49:35.835999 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:49:35 crc kubenswrapper[4706]: I1208 19:49:35.836829 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:50:05 crc kubenswrapper[4706]: I1208 19:50:05.836853 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:50:05 crc kubenswrapper[4706]: I1208 19:50:05.837749 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:50:05 crc kubenswrapper[4706]: I1208 19:50:05.837827 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:50:05 crc kubenswrapper[4706]: I1208 19:50:05.839002 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:50:05 crc kubenswrapper[4706]: I1208 19:50:05.839063 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" gracePeriod=600 Dec 08 19:50:05 crc kubenswrapper[4706]: E1208 19:50:05.972526 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:50:06 crc kubenswrapper[4706]: I1208 19:50:06.282876 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" exitCode=0 Dec 08 19:50:06 crc kubenswrapper[4706]: I1208 19:50:06.282931 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494"} Dec 08 19:50:06 crc kubenswrapper[4706]: I1208 19:50:06.282976 4706 scope.go:117] "RemoveContainer" containerID="a5681b5fe1d33f6e2441d03b7699cdb0e62dc0c63f422690f0cadd789557d42f" Dec 08 19:50:06 crc kubenswrapper[4706]: I1208 19:50:06.284165 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:50:06 crc kubenswrapper[4706]: E1208 19:50:06.284607 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:50:17 crc kubenswrapper[4706]: I1208 19:50:17.617093 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:50:17 crc kubenswrapper[4706]: E1208 19:50:17.618449 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:50:28 crc kubenswrapper[4706]: I1208 19:50:28.610201 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:50:28 crc kubenswrapper[4706]: E1208 19:50:28.611427 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:50:41 crc kubenswrapper[4706]: I1208 19:50:41.612247 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:50:41 crc kubenswrapper[4706]: E1208 19:50:41.613657 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:50:56 crc kubenswrapper[4706]: I1208 19:50:56.609854 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:50:56 crc kubenswrapper[4706]: E1208 19:50:56.610955 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:51:07 crc kubenswrapper[4706]: I1208 19:51:07.616321 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:51:07 crc kubenswrapper[4706]: E1208 19:51:07.617618 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:51:20 crc kubenswrapper[4706]: I1208 19:51:20.055971 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c9ab-account-create-update-b72xz"] Dec 08 19:51:20 crc kubenswrapper[4706]: I1208 19:51:20.071779 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-pvxvl"] Dec 08 19:51:20 crc kubenswrapper[4706]: I1208 19:51:20.083685 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c9ab-account-create-update-b72xz"] Dec 08 19:51:20 crc kubenswrapper[4706]: I1208 19:51:20.094384 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-pvxvl"] Dec 08 19:51:20 crc kubenswrapper[4706]: I1208 19:51:20.609600 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:51:20 crc kubenswrapper[4706]: E1208 19:51:20.610045 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:51:21 crc kubenswrapper[4706]: I1208 19:51:21.622077 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a36b48c2-0389-416b-a142-b28dc1ea76da" path="/var/lib/kubelet/pods/a36b48c2-0389-416b-a142-b28dc1ea76da/volumes" Dec 08 19:51:21 crc kubenswrapper[4706]: I1208 19:51:21.623112 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce25e3ff-7e7a-4776-a1fb-bec77f08fa56" path="/var/lib/kubelet/pods/ce25e3ff-7e7a-4776-a1fb-bec77f08fa56/volumes" Dec 08 19:51:24 crc kubenswrapper[4706]: I1208 19:51:24.073286 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-szsrk"] Dec 08 19:51:24 crc kubenswrapper[4706]: I1208 19:51:24.085731 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec9a-account-create-update-bfpsl"] Dec 08 19:51:24 crc kubenswrapper[4706]: I1208 19:51:24.097852 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-szsrk"] Dec 08 19:51:24 crc kubenswrapper[4706]: I1208 19:51:24.109006 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ec9a-account-create-update-bfpsl"] Dec 08 19:51:25 crc kubenswrapper[4706]: I1208 19:51:25.621818 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091f4a40-69b6-4c09-8251-d5aca75a420f" path="/var/lib/kubelet/pods/091f4a40-69b6-4c09-8251-d5aca75a420f/volumes" Dec 08 19:51:25 crc kubenswrapper[4706]: I1208 19:51:25.623450 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f23ba90a-3542-4241-9e27-63a9bcd2baee" path="/var/lib/kubelet/pods/f23ba90a-3542-4241-9e27-63a9bcd2baee/volumes" Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.041658 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-create-6lf8v"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.058578 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b7de-account-create-update-ntz9l"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.075989 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-cpmsk"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.089923 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-cpmsk"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.102806 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-create-6lf8v"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.118294 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-dfae-account-create-update-dhrlq"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.129997 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b7de-account-create-update-ntz9l"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.141680 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-xlsrb"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.153910 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-dfae-account-create-update-dhrlq"] Dec 08 19:51:26 crc kubenswrapper[4706]: I1208 19:51:26.166462 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-xlsrb"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.038288 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-3e4b-account-create-update-ghzkt"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.051032 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-pjffv"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.065315 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2e9f-account-create-update-9kkfk"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.080344 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-3e4b-account-create-update-ghzkt"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.091115 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-04cd-account-create-update-zbfff"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.105397 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-248t7"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.127008 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2e9f-account-create-update-9kkfk"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.142953 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-pjffv"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.183905 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-248t7"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.203314 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-04cd-account-create-update-zbfff"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.572951 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4xg64"] Dec 08 19:51:27 crc kubenswrapper[4706]: E1208 19:51:27.573742 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" containerName="registry-server" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.573768 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" containerName="registry-server" Dec 08 19:51:27 crc kubenswrapper[4706]: E1208 19:51:27.573786 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" containerName="extract-content" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.573795 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" containerName="extract-content" Dec 08 19:51:27 crc kubenswrapper[4706]: E1208 19:51:27.573835 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" containerName="extract-utilities" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.573845 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" containerName="extract-utilities" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.574196 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c91542a-001c-48aa-9caf-94d2a35af371" containerName="registry-server" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.577866 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.589052 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xg64"] Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.634819 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f81f35c-cf94-4180-ad69-3d54f8c41a35" path="/var/lib/kubelet/pods/3f81f35c-cf94-4180-ad69-3d54f8c41a35/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.635907 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50a60ccd-c961-4c0c-ae74-0b0d9c6e172a" path="/var/lib/kubelet/pods/50a60ccd-c961-4c0c-ae74-0b0d9c6e172a/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.636532 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d89be3-c6e5-4f3e-b788-4811b47ba332" path="/var/lib/kubelet/pods/53d89be3-c6e5-4f3e-b788-4811b47ba332/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.637107 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a7451c2-8247-406f-9d8f-eb5690a16555" path="/var/lib/kubelet/pods/5a7451c2-8247-406f-9d8f-eb5690a16555/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.638227 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6507d8ce-f566-46d6-b55d-abbaccc71857" path="/var/lib/kubelet/pods/6507d8ce-f566-46d6-b55d-abbaccc71857/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.638867 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d1a2245-a7e5-4f5f-a392-da5c9aec368a" path="/var/lib/kubelet/pods/6d1a2245-a7e5-4f5f-a392-da5c9aec368a/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.639527 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2583ff3-e358-446c-9660-db7074583d8a" path="/var/lib/kubelet/pods/a2583ff3-e358-446c-9660-db7074583d8a/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.640612 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb919dfd-81a4-42b1-8e30-4788113753d2" path="/var/lib/kubelet/pods/bb919dfd-81a4-42b1-8e30-4788113753d2/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.641176 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1bc7a6a-8c78-4597-ba47-23e0270264e4" path="/var/lib/kubelet/pods/e1bc7a6a-8c78-4597-ba47-23e0270264e4/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.641819 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd2c1239-cb7e-4bc1-af35-75fcd1613c77" path="/var/lib/kubelet/pods/fd2c1239-cb7e-4bc1-af35-75fcd1613c77/volumes" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.725526 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whk9w\" (UniqueName: \"kubernetes.io/projected/2024864b-a9b0-474e-ad3c-346b394bc0a3-kube-api-access-whk9w\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.725758 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-catalog-content\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.726086 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-utilities\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.828940 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-utilities\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.829073 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whk9w\" (UniqueName: \"kubernetes.io/projected/2024864b-a9b0-474e-ad3c-346b394bc0a3-kube-api-access-whk9w\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.829157 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-catalog-content\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.830229 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-catalog-content\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.830540 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-utilities\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.858090 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whk9w\" (UniqueName: \"kubernetes.io/projected/2024864b-a9b0-474e-ad3c-346b394bc0a3-kube-api-access-whk9w\") pod \"redhat-operators-4xg64\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:27 crc kubenswrapper[4706]: I1208 19:51:27.905781 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:28 crc kubenswrapper[4706]: I1208 19:51:28.515198 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xg64"] Dec 08 19:51:29 crc kubenswrapper[4706]: I1208 19:51:29.367845 4706 generic.go:334] "Generic (PLEG): container finished" podID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerID="801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd" exitCode=0 Dec 08 19:51:29 crc kubenswrapper[4706]: I1208 19:51:29.367924 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xg64" event={"ID":"2024864b-a9b0-474e-ad3c-346b394bc0a3","Type":"ContainerDied","Data":"801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd"} Dec 08 19:51:29 crc kubenswrapper[4706]: I1208 19:51:29.368358 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xg64" event={"ID":"2024864b-a9b0-474e-ad3c-346b394bc0a3","Type":"ContainerStarted","Data":"435cc3d87808a9d2283f1ebf49f9766013c7a0cd82e38b3233b6a6859eb96ff5"} Dec 08 19:51:30 crc kubenswrapper[4706]: I1208 19:51:30.381215 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xg64" event={"ID":"2024864b-a9b0-474e-ad3c-346b394bc0a3","Type":"ContainerStarted","Data":"4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c"} Dec 08 19:51:31 crc kubenswrapper[4706]: I1208 19:51:31.609238 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:51:31 crc kubenswrapper[4706]: E1208 19:51:31.609938 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:51:34 crc kubenswrapper[4706]: I1208 19:51:34.436983 4706 generic.go:334] "Generic (PLEG): container finished" podID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerID="4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c" exitCode=0 Dec 08 19:51:34 crc kubenswrapper[4706]: I1208 19:51:34.437097 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xg64" event={"ID":"2024864b-a9b0-474e-ad3c-346b394bc0a3","Type":"ContainerDied","Data":"4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c"} Dec 08 19:51:35 crc kubenswrapper[4706]: I1208 19:51:35.450732 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xg64" event={"ID":"2024864b-a9b0-474e-ad3c-346b394bc0a3","Type":"ContainerStarted","Data":"1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91"} Dec 08 19:51:35 crc kubenswrapper[4706]: I1208 19:51:35.478001 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4xg64" podStartSLOduration=2.947445482 podStartE2EDuration="8.477977345s" podCreationTimestamp="2025-12-08 19:51:27 +0000 UTC" firstStartedPulling="2025-12-08 19:51:29.37018701 +0000 UTC m=+1772.012388013" lastFinishedPulling="2025-12-08 19:51:34.900718873 +0000 UTC m=+1777.542919876" observedRunningTime="2025-12-08 19:51:35.471255421 +0000 UTC m=+1778.113456444" watchObservedRunningTime="2025-12-08 19:51:35.477977345 +0000 UTC m=+1778.120178348" Dec 08 19:51:37 crc kubenswrapper[4706]: I1208 19:51:37.906940 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:37 crc kubenswrapper[4706]: I1208 19:51:37.907284 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:38 crc kubenswrapper[4706]: I1208 19:51:38.956671 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4xg64" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="registry-server" probeResult="failure" output=< Dec 08 19:51:38 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 19:51:38 crc kubenswrapper[4706]: > Dec 08 19:51:44 crc kubenswrapper[4706]: I1208 19:51:44.609482 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:51:44 crc kubenswrapper[4706]: E1208 19:51:44.610694 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:51:47 crc kubenswrapper[4706]: I1208 19:51:47.970981 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:48 crc kubenswrapper[4706]: I1208 19:51:48.024510 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:48 crc kubenswrapper[4706]: I1208 19:51:48.219432 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xg64"] Dec 08 19:51:49 crc kubenswrapper[4706]: I1208 19:51:49.604542 4706 generic.go:334] "Generic (PLEG): container finished" podID="91cd912d-3492-4709-99e9-94fa496bde14" containerID="63bfb742e3057571684fdd52cbe981af33b67db6bbdf7e574dfab7a8bddb4206" exitCode=0 Dec 08 19:51:49 crc kubenswrapper[4706]: I1208 19:51:49.604697 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" event={"ID":"91cd912d-3492-4709-99e9-94fa496bde14","Type":"ContainerDied","Data":"63bfb742e3057571684fdd52cbe981af33b67db6bbdf7e574dfab7a8bddb4206"} Dec 08 19:51:49 crc kubenswrapper[4706]: I1208 19:51:49.605435 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4xg64" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="registry-server" containerID="cri-o://1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91" gracePeriod=2 Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.239451 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.313859 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-utilities\") pod \"2024864b-a9b0-474e-ad3c-346b394bc0a3\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.314233 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-catalog-content\") pod \"2024864b-a9b0-474e-ad3c-346b394bc0a3\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.314440 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whk9w\" (UniqueName: \"kubernetes.io/projected/2024864b-a9b0-474e-ad3c-346b394bc0a3-kube-api-access-whk9w\") pod \"2024864b-a9b0-474e-ad3c-346b394bc0a3\" (UID: \"2024864b-a9b0-474e-ad3c-346b394bc0a3\") " Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.315106 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-utilities" (OuterVolumeSpecName: "utilities") pod "2024864b-a9b0-474e-ad3c-346b394bc0a3" (UID: "2024864b-a9b0-474e-ad3c-346b394bc0a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.323609 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2024864b-a9b0-474e-ad3c-346b394bc0a3-kube-api-access-whk9w" (OuterVolumeSpecName: "kube-api-access-whk9w") pod "2024864b-a9b0-474e-ad3c-346b394bc0a3" (UID: "2024864b-a9b0-474e-ad3c-346b394bc0a3"). InnerVolumeSpecName "kube-api-access-whk9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.419366 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whk9w\" (UniqueName: \"kubernetes.io/projected/2024864b-a9b0-474e-ad3c-346b394bc0a3-kube-api-access-whk9w\") on node \"crc\" DevicePath \"\"" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.419410 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.433523 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2024864b-a9b0-474e-ad3c-346b394bc0a3" (UID: "2024864b-a9b0-474e-ad3c-346b394bc0a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.522924 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2024864b-a9b0-474e-ad3c-346b394bc0a3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.621040 4706 generic.go:334] "Generic (PLEG): container finished" podID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerID="1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91" exitCode=0 Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.621103 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xg64" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.621157 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xg64" event={"ID":"2024864b-a9b0-474e-ad3c-346b394bc0a3","Type":"ContainerDied","Data":"1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91"} Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.621225 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xg64" event={"ID":"2024864b-a9b0-474e-ad3c-346b394bc0a3","Type":"ContainerDied","Data":"435cc3d87808a9d2283f1ebf49f9766013c7a0cd82e38b3233b6a6859eb96ff5"} Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.621329 4706 scope.go:117] "RemoveContainer" containerID="1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.660765 4706 scope.go:117] "RemoveContainer" containerID="4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.688670 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xg64"] Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.702025 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4xg64"] Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.716591 4706 scope.go:117] "RemoveContainer" containerID="801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.776639 4706 scope.go:117] "RemoveContainer" containerID="1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91" Dec 08 19:51:50 crc kubenswrapper[4706]: E1208 19:51:50.777898 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91\": container with ID starting with 1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91 not found: ID does not exist" containerID="1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.777939 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91"} err="failed to get container status \"1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91\": rpc error: code = NotFound desc = could not find container \"1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91\": container with ID starting with 1b4f31cd2609265e3c22cb0471da939ee6263f82fa4ee1341ab9be9183f01b91 not found: ID does not exist" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.777964 4706 scope.go:117] "RemoveContainer" containerID="4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c" Dec 08 19:51:50 crc kubenswrapper[4706]: E1208 19:51:50.779064 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c\": container with ID starting with 4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c not found: ID does not exist" containerID="4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.779116 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c"} err="failed to get container status \"4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c\": rpc error: code = NotFound desc = could not find container \"4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c\": container with ID starting with 4981c87ffe2d29711f9db64f324c61072c1518aecf444512655b62bb781ed53c not found: ID does not exist" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.779153 4706 scope.go:117] "RemoveContainer" containerID="801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd" Dec 08 19:51:50 crc kubenswrapper[4706]: E1208 19:51:50.782876 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd\": container with ID starting with 801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd not found: ID does not exist" containerID="801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd" Dec 08 19:51:50 crc kubenswrapper[4706]: I1208 19:51:50.782915 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd"} err="failed to get container status \"801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd\": rpc error: code = NotFound desc = could not find container \"801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd\": container with ID starting with 801d59819feccee9ede714e92269fb77d8af7addc0a3ccdf90ae49c24abe4dbd not found: ID does not exist" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.246487 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.373732 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-inventory\") pod \"91cd912d-3492-4709-99e9-94fa496bde14\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.373834 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-bootstrap-combined-ca-bundle\") pod \"91cd912d-3492-4709-99e9-94fa496bde14\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.374049 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-ssh-key\") pod \"91cd912d-3492-4709-99e9-94fa496bde14\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.374180 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk6gd\" (UniqueName: \"kubernetes.io/projected/91cd912d-3492-4709-99e9-94fa496bde14-kube-api-access-lk6gd\") pod \"91cd912d-3492-4709-99e9-94fa496bde14\" (UID: \"91cd912d-3492-4709-99e9-94fa496bde14\") " Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.383647 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91cd912d-3492-4709-99e9-94fa496bde14-kube-api-access-lk6gd" (OuterVolumeSpecName: "kube-api-access-lk6gd") pod "91cd912d-3492-4709-99e9-94fa496bde14" (UID: "91cd912d-3492-4709-99e9-94fa496bde14"). InnerVolumeSpecName "kube-api-access-lk6gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.384042 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "91cd912d-3492-4709-99e9-94fa496bde14" (UID: "91cd912d-3492-4709-99e9-94fa496bde14"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.411704 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-inventory" (OuterVolumeSpecName: "inventory") pod "91cd912d-3492-4709-99e9-94fa496bde14" (UID: "91cd912d-3492-4709-99e9-94fa496bde14"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.422521 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "91cd912d-3492-4709-99e9-94fa496bde14" (UID: "91cd912d-3492-4709-99e9-94fa496bde14"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.481648 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.481683 4706 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.481698 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91cd912d-3492-4709-99e9-94fa496bde14-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.481708 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk6gd\" (UniqueName: \"kubernetes.io/projected/91cd912d-3492-4709-99e9-94fa496bde14-kube-api-access-lk6gd\") on node \"crc\" DevicePath \"\"" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.625854 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" path="/var/lib/kubelet/pods/2024864b-a9b0-474e-ad3c-346b394bc0a3/volumes" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.640438 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" event={"ID":"91cd912d-3492-4709-99e9-94fa496bde14","Type":"ContainerDied","Data":"9c227434af50512f5071448c3f00a2b28b56966f7f94e154f6286088ba52ea82"} Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.640753 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c227434af50512f5071448c3f00a2b28b56966f7f94e154f6286088ba52ea82" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.640474 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-267vl" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.725166 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s"] Dec 08 19:51:51 crc kubenswrapper[4706]: E1208 19:51:51.725752 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="extract-content" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.725780 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="extract-content" Dec 08 19:51:51 crc kubenswrapper[4706]: E1208 19:51:51.725806 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91cd912d-3492-4709-99e9-94fa496bde14" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.725818 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="91cd912d-3492-4709-99e9-94fa496bde14" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 08 19:51:51 crc kubenswrapper[4706]: E1208 19:51:51.725851 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="registry-server" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.725859 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="registry-server" Dec 08 19:51:51 crc kubenswrapper[4706]: E1208 19:51:51.725881 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="extract-utilities" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.725916 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="extract-utilities" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.726143 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="91cd912d-3492-4709-99e9-94fa496bde14" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.726171 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2024864b-a9b0-474e-ad3c-346b394bc0a3" containerName="registry-server" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.727456 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.734835 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.734972 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.734848 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.735319 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.749026 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s"] Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.894998 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.895607 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.895812 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrx9j\" (UniqueName: \"kubernetes.io/projected/a9260164-eeea-47a2-9502-1744cba85586-kube-api-access-zrx9j\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.998246 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.998313 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrx9j\" (UniqueName: \"kubernetes.io/projected/a9260164-eeea-47a2-9502-1744cba85586-kube-api-access-zrx9j\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:51 crc kubenswrapper[4706]: I1208 19:51:51.998371 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:52 crc kubenswrapper[4706]: I1208 19:51:52.004149 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:52 crc kubenswrapper[4706]: I1208 19:51:52.004149 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:52 crc kubenswrapper[4706]: I1208 19:51:52.024797 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrx9j\" (UniqueName: \"kubernetes.io/projected/a9260164-eeea-47a2-9502-1744cba85586-kube-api-access-zrx9j\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:52 crc kubenswrapper[4706]: I1208 19:51:52.047721 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:51:52 crc kubenswrapper[4706]: I1208 19:51:52.066952 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-g8g9r"] Dec 08 19:51:52 crc kubenswrapper[4706]: I1208 19:51:52.078991 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-g8g9r"] Dec 08 19:51:52 crc kubenswrapper[4706]: I1208 19:51:52.614191 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s"] Dec 08 19:51:52 crc kubenswrapper[4706]: I1208 19:51:52.659721 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" event={"ID":"a9260164-eeea-47a2-9502-1744cba85586","Type":"ContainerStarted","Data":"b485a0498c03604abb4dbf0b59c8a59ad50d6a0f81a25d59768cbf264deea75e"} Dec 08 19:51:53 crc kubenswrapper[4706]: I1208 19:51:53.624509 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b499661-3a56-4b32-ada0-d0a132752928" path="/var/lib/kubelet/pods/3b499661-3a56-4b32-ada0-d0a132752928/volumes" Dec 08 19:51:53 crc kubenswrapper[4706]: I1208 19:51:53.678004 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" event={"ID":"a9260164-eeea-47a2-9502-1744cba85586","Type":"ContainerStarted","Data":"4d4e1747cc5b334821f1d6fe80d08b33b86e02198a1233ceb52dcc1190ac84ef"} Dec 08 19:51:53 crc kubenswrapper[4706]: I1208 19:51:53.710633 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" podStartSLOduration=2.186737924 podStartE2EDuration="2.71060693s" podCreationTimestamp="2025-12-08 19:51:51 +0000 UTC" firstStartedPulling="2025-12-08 19:51:52.618938393 +0000 UTC m=+1795.261139396" lastFinishedPulling="2025-12-08 19:51:53.142807409 +0000 UTC m=+1795.785008402" observedRunningTime="2025-12-08 19:51:53.696521415 +0000 UTC m=+1796.338722418" watchObservedRunningTime="2025-12-08 19:51:53.71060693 +0000 UTC m=+1796.352807923" Dec 08 19:51:58 crc kubenswrapper[4706]: I1208 19:51:58.609058 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:51:58 crc kubenswrapper[4706]: E1208 19:51:58.610086 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:52:09 crc kubenswrapper[4706]: I1208 19:52:09.608671 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:52:09 crc kubenswrapper[4706]: E1208 19:52:09.609668 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.115148 4706 scope.go:117] "RemoveContainer" containerID="286bb523070edb6a5131de79c1e7d90a4a41362b076e9aa5d5b6b2a0e851e984" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.161756 4706 scope.go:117] "RemoveContainer" containerID="600d5def7d38461ce26b35d9038face2cc040047cebb58b0ade2fefcaad591ef" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.242822 4706 scope.go:117] "RemoveContainer" containerID="410357fb3263e326acd30036fd69bcf7fe0c2b8a434de30b3506236433235dba" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.291339 4706 scope.go:117] "RemoveContainer" containerID="af4da9490fac94bb822be0a586032bbf3570469c690e73984ee6d116c975e496" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.340485 4706 scope.go:117] "RemoveContainer" containerID="9a3436d1e435ccfa19264d07fdaa4e33e7ef3a978b4748d07a75554255977370" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.417503 4706 scope.go:117] "RemoveContainer" containerID="1e35ee43159ba60bac417a2197bb3ac0f6c9d3ca476a87ce1526580a19b706da" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.479075 4706 scope.go:117] "RemoveContainer" containerID="052441d1f5ffa04ff8fa2acadb9a0fb45905f41303dfb139d97ec38a230b3b82" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.516981 4706 scope.go:117] "RemoveContainer" containerID="17c235c39e4638c7623d7a98b4579a281bb377c5ce7a151855daa8e97f18c37e" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.542326 4706 scope.go:117] "RemoveContainer" containerID="89fd8dc0f0a6ef5e3f3f07e8d9cdefed1911f9274b9723c5165549b21a5a9e77" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.569084 4706 scope.go:117] "RemoveContainer" containerID="481f9f540680ba35cb6cbf21b26f617bc25d9d5fce4bba116507a1ea39fcd8d4" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.597171 4706 scope.go:117] "RemoveContainer" containerID="f8fb649ccf58b41a0af70194c43aff3468d21ddcad19af964b7e4496056ed6dc" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.622122 4706 scope.go:117] "RemoveContainer" containerID="43e0453592947499e12f422ebc2fc68b6ddf6320758593f1867e1a82293cc3c6" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.649475 4706 scope.go:117] "RemoveContainer" containerID="f14ffbe72b87da1bf06aed478ace4a9bd69b5cd72f7321aa5d834315bfcbdb29" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.682862 4706 scope.go:117] "RemoveContainer" containerID="03710859b1f1f8aede28d1c0570fe1e5ee1b7eb978dd4f839d9e220d609766bc" Dec 08 19:52:10 crc kubenswrapper[4706]: I1208 19:52:10.709697 4706 scope.go:117] "RemoveContainer" containerID="2796983a054cd5070dabaa79457a6d5db52c03892764167985d2edb3a6242898" Dec 08 19:52:20 crc kubenswrapper[4706]: I1208 19:52:20.609119 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:52:20 crc kubenswrapper[4706]: E1208 19:52:20.610285 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:52:31 crc kubenswrapper[4706]: I1208 19:52:31.610316 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:52:31 crc kubenswrapper[4706]: E1208 19:52:31.611502 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:52:32 crc kubenswrapper[4706]: I1208 19:52:32.060660 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-dkmkk"] Dec 08 19:52:32 crc kubenswrapper[4706]: I1208 19:52:32.076662 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-dkmkk"] Dec 08 19:52:33 crc kubenswrapper[4706]: I1208 19:52:33.629172 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18ef7995-14d2-4c7a-8b30-8aaac1c49b92" path="/var/lib/kubelet/pods/18ef7995-14d2-4c7a-8b30-8aaac1c49b92/volumes" Dec 08 19:52:44 crc kubenswrapper[4706]: I1208 19:52:44.608779 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:52:44 crc kubenswrapper[4706]: E1208 19:52:44.610111 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:52:50 crc kubenswrapper[4706]: I1208 19:52:50.047831 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-lhc9c"] Dec 08 19:52:50 crc kubenswrapper[4706]: I1208 19:52:50.062221 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ff9kc"] Dec 08 19:52:50 crc kubenswrapper[4706]: I1208 19:52:50.077062 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jhpwq"] Dec 08 19:52:50 crc kubenswrapper[4706]: I1208 19:52:50.090758 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-lhc9c"] Dec 08 19:52:50 crc kubenswrapper[4706]: I1208 19:52:50.103849 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ff9kc"] Dec 08 19:52:50 crc kubenswrapper[4706]: I1208 19:52:50.116721 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jhpwq"] Dec 08 19:52:51 crc kubenswrapper[4706]: I1208 19:52:51.624326 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8677c722-159f-4509-a1b1-114817ef241b" path="/var/lib/kubelet/pods/8677c722-159f-4509-a1b1-114817ef241b/volumes" Dec 08 19:52:51 crc kubenswrapper[4706]: I1208 19:52:51.625062 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d" path="/var/lib/kubelet/pods/c2dc07a4-5f58-43b2-918c-ea0a5fbb2b6d/volumes" Dec 08 19:52:51 crc kubenswrapper[4706]: I1208 19:52:51.627761 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6007166-5c77-46c7-8c2a-e4ff780725d6" path="/var/lib/kubelet/pods/d6007166-5c77-46c7-8c2a-e4ff780725d6/volumes" Dec 08 19:52:59 crc kubenswrapper[4706]: I1208 19:52:59.610228 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:52:59 crc kubenswrapper[4706]: E1208 19:52:59.612622 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:53:05 crc kubenswrapper[4706]: I1208 19:53:05.036296 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-8s7bd"] Dec 08 19:53:05 crc kubenswrapper[4706]: I1208 19:53:05.046309 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-8s7bd"] Dec 08 19:53:05 crc kubenswrapper[4706]: I1208 19:53:05.624558 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e18991e1-84d5-4e32-9a05-ad08f31cbd60" path="/var/lib/kubelet/pods/e18991e1-84d5-4e32-9a05-ad08f31cbd60/volumes" Dec 08 19:53:06 crc kubenswrapper[4706]: I1208 19:53:06.043376 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-brz4k"] Dec 08 19:53:06 crc kubenswrapper[4706]: I1208 19:53:06.061347 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-brz4k"] Dec 08 19:53:07 crc kubenswrapper[4706]: I1208 19:53:07.623944 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="776cb4c2-52c5-404a-b3ca-c83f93dcbd0e" path="/var/lib/kubelet/pods/776cb4c2-52c5-404a-b3ca-c83f93dcbd0e/volumes" Dec 08 19:53:11 crc kubenswrapper[4706]: I1208 19:53:11.091603 4706 scope.go:117] "RemoveContainer" containerID="c9032f02ec92f5cce032c294f93e9863342b56a15ee19e00d84c0f13dd8145e1" Dec 08 19:53:11 crc kubenswrapper[4706]: I1208 19:53:11.139889 4706 scope.go:117] "RemoveContainer" containerID="843d7a7fa9e2d4430c3baedec98c6e57411f55a0075a2d7e5799758eaadae500" Dec 08 19:53:11 crc kubenswrapper[4706]: I1208 19:53:11.214362 4706 scope.go:117] "RemoveContainer" containerID="75ea842913c75d2f26e428dc0a0703205af3a7cea0a7ce82fee7e017f69ca031" Dec 08 19:53:11 crc kubenswrapper[4706]: I1208 19:53:11.272352 4706 scope.go:117] "RemoveContainer" containerID="62ff1d077e6f82a9f25fc066972b37cf8127ed601d34a18acbb4ddfb98445689" Dec 08 19:53:11 crc kubenswrapper[4706]: I1208 19:53:11.316874 4706 scope.go:117] "RemoveContainer" containerID="c82b5de45969cfe9bdd14bd26c0cf70ebdc1ca0cf543a3250b7973a433f31d1c" Dec 08 19:53:11 crc kubenswrapper[4706]: I1208 19:53:11.353069 4706 scope.go:117] "RemoveContainer" containerID="bb6f1a9e15babd2f25e5b3f17b0c5392aa16c1ca08b9fc1aeac629c0e3e9de33" Dec 08 19:53:11 crc kubenswrapper[4706]: I1208 19:53:11.609934 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:53:11 crc kubenswrapper[4706]: E1208 19:53:11.610577 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:53:23 crc kubenswrapper[4706]: I1208 19:53:23.609685 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:53:23 crc kubenswrapper[4706]: E1208 19:53:23.610696 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:53:36 crc kubenswrapper[4706]: I1208 19:53:36.608294 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:53:36 crc kubenswrapper[4706]: E1208 19:53:36.609254 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:53:50 crc kubenswrapper[4706]: I1208 19:53:50.608572 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:53:50 crc kubenswrapper[4706]: E1208 19:53:50.610946 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:54:03 crc kubenswrapper[4706]: I1208 19:54:03.224052 4706 generic.go:334] "Generic (PLEG): container finished" podID="a9260164-eeea-47a2-9502-1744cba85586" containerID="4d4e1747cc5b334821f1d6fe80d08b33b86e02198a1233ceb52dcc1190ac84ef" exitCode=0 Dec 08 19:54:03 crc kubenswrapper[4706]: I1208 19:54:03.224175 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" event={"ID":"a9260164-eeea-47a2-9502-1744cba85586","Type":"ContainerDied","Data":"4d4e1747cc5b334821f1d6fe80d08b33b86e02198a1233ceb52dcc1190ac84ef"} Dec 08 19:54:03 crc kubenswrapper[4706]: I1208 19:54:03.611016 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:54:03 crc kubenswrapper[4706]: E1208 19:54:03.611734 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:54:04 crc kubenswrapper[4706]: I1208 19:54:04.808733 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:54:04 crc kubenswrapper[4706]: I1208 19:54:04.973540 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-inventory\") pod \"a9260164-eeea-47a2-9502-1744cba85586\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " Dec 08 19:54:04 crc kubenswrapper[4706]: I1208 19:54:04.973779 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrx9j\" (UniqueName: \"kubernetes.io/projected/a9260164-eeea-47a2-9502-1744cba85586-kube-api-access-zrx9j\") pod \"a9260164-eeea-47a2-9502-1744cba85586\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " Dec 08 19:54:04 crc kubenswrapper[4706]: I1208 19:54:04.974553 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-ssh-key\") pod \"a9260164-eeea-47a2-9502-1744cba85586\" (UID: \"a9260164-eeea-47a2-9502-1744cba85586\") " Dec 08 19:54:04 crc kubenswrapper[4706]: I1208 19:54:04.991399 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9260164-eeea-47a2-9502-1744cba85586-kube-api-access-zrx9j" (OuterVolumeSpecName: "kube-api-access-zrx9j") pod "a9260164-eeea-47a2-9502-1744cba85586" (UID: "a9260164-eeea-47a2-9502-1744cba85586"). InnerVolumeSpecName "kube-api-access-zrx9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.011388 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-inventory" (OuterVolumeSpecName: "inventory") pod "a9260164-eeea-47a2-9502-1744cba85586" (UID: "a9260164-eeea-47a2-9502-1744cba85586"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.013575 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a9260164-eeea-47a2-9502-1744cba85586" (UID: "a9260164-eeea-47a2-9502-1744cba85586"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.078379 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.078411 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrx9j\" (UniqueName: \"kubernetes.io/projected/a9260164-eeea-47a2-9502-1744cba85586-kube-api-access-zrx9j\") on node \"crc\" DevicePath \"\"" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.078423 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9260164-eeea-47a2-9502-1744cba85586-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.253825 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" event={"ID":"a9260164-eeea-47a2-9502-1744cba85586","Type":"ContainerDied","Data":"b485a0498c03604abb4dbf0b59c8a59ad50d6a0f81a25d59768cbf264deea75e"} Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.253884 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b485a0498c03604abb4dbf0b59c8a59ad50d6a0f81a25d59768cbf264deea75e" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.253933 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.360103 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx"] Dec 08 19:54:05 crc kubenswrapper[4706]: E1208 19:54:05.361604 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9260164-eeea-47a2-9502-1744cba85586" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.361746 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9260164-eeea-47a2-9502-1744cba85586" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.364132 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9260164-eeea-47a2-9502-1744cba85586" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.365421 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.372204 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.372741 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.372949 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.377781 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.383056 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx"] Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.488550 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bdrr\" (UniqueName: \"kubernetes.io/projected/784135f7-ff54-43df-b7e7-ea4f0b940fdc-kube-api-access-4bdrr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.489094 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.489438 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.592222 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.592917 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bdrr\" (UniqueName: \"kubernetes.io/projected/784135f7-ff54-43df-b7e7-ea4f0b940fdc-kube-api-access-4bdrr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.593299 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.598890 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.609973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.612079 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bdrr\" (UniqueName: \"kubernetes.io/projected/784135f7-ff54-43df-b7e7-ea4f0b940fdc-kube-api-access-4bdrr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:05 crc kubenswrapper[4706]: I1208 19:54:05.698226 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:54:06 crc kubenswrapper[4706]: I1208 19:54:06.322928 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx"] Dec 08 19:54:06 crc kubenswrapper[4706]: I1208 19:54:06.336069 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 19:54:07 crc kubenswrapper[4706]: I1208 19:54:07.282021 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" event={"ID":"784135f7-ff54-43df-b7e7-ea4f0b940fdc","Type":"ContainerStarted","Data":"73e8ef0a119a93b618644622461dba2ef2d842900995883256490370eb44de6f"} Dec 08 19:54:07 crc kubenswrapper[4706]: I1208 19:54:07.282910 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" event={"ID":"784135f7-ff54-43df-b7e7-ea4f0b940fdc","Type":"ContainerStarted","Data":"3b50c26b0d9565c688ebd8600a0508245875ed4af6d99b355c5c147c9e2b7382"} Dec 08 19:54:07 crc kubenswrapper[4706]: I1208 19:54:07.307646 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" podStartSLOduration=1.8582368580000002 podStartE2EDuration="2.30761546s" podCreationTimestamp="2025-12-08 19:54:05 +0000 UTC" firstStartedPulling="2025-12-08 19:54:06.335562402 +0000 UTC m=+1928.977763405" lastFinishedPulling="2025-12-08 19:54:06.784941004 +0000 UTC m=+1929.427142007" observedRunningTime="2025-12-08 19:54:07.303227815 +0000 UTC m=+1929.945428818" watchObservedRunningTime="2025-12-08 19:54:07.30761546 +0000 UTC m=+1929.949816473" Dec 08 19:54:15 crc kubenswrapper[4706]: I1208 19:54:15.608916 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:54:15 crc kubenswrapper[4706]: E1208 19:54:15.610070 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:54:16 crc kubenswrapper[4706]: I1208 19:54:16.088637 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-kd4rj"] Dec 08 19:54:16 crc kubenswrapper[4706]: I1208 19:54:16.104020 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-t86b4"] Dec 08 19:54:16 crc kubenswrapper[4706]: I1208 19:54:16.115007 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-c82mx"] Dec 08 19:54:16 crc kubenswrapper[4706]: I1208 19:54:16.124781 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-758b-account-create-update-zh9x9"] Dec 08 19:54:16 crc kubenswrapper[4706]: I1208 19:54:16.135075 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-t86b4"] Dec 08 19:54:16 crc kubenswrapper[4706]: I1208 19:54:16.148755 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-c82mx"] Dec 08 19:54:16 crc kubenswrapper[4706]: I1208 19:54:16.160737 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-kd4rj"] Dec 08 19:54:16 crc kubenswrapper[4706]: I1208 19:54:16.171169 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-758b-account-create-update-zh9x9"] Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.039161 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2143-account-create-update-fqwzd"] Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.055511 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c56a-account-create-update-kl49q"] Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.068707 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2143-account-create-update-fqwzd"] Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.078535 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c56a-account-create-update-kl49q"] Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.643424 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ffa1493-bb68-48b6-8f13-c9c30358d1f3" path="/var/lib/kubelet/pods/3ffa1493-bb68-48b6-8f13-c9c30358d1f3/volumes" Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.646796 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50786407-6513-4d04-b13b-0367252b90af" path="/var/lib/kubelet/pods/50786407-6513-4d04-b13b-0367252b90af/volumes" Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.647739 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66571791-d210-477a-8951-20a0609193a9" path="/var/lib/kubelet/pods/66571791-d210-477a-8951-20a0609193a9/volumes" Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.648622 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="855d2f91-5dfa-401f-b1ac-a3f94287c245" path="/var/lib/kubelet/pods/855d2f91-5dfa-401f-b1ac-a3f94287c245/volumes" Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.650434 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="868f8ce6-c1f8-404d-8226-7fffb0518de4" path="/var/lib/kubelet/pods/868f8ce6-c1f8-404d-8226-7fffb0518de4/volumes" Dec 08 19:54:17 crc kubenswrapper[4706]: I1208 19:54:17.652156 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05ed39a-048e-4de0-812c-99daca774a99" path="/var/lib/kubelet/pods/d05ed39a-048e-4de0-812c-99daca774a99/volumes" Dec 08 19:54:26 crc kubenswrapper[4706]: I1208 19:54:26.610352 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:54:26 crc kubenswrapper[4706]: E1208 19:54:26.611232 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:54:40 crc kubenswrapper[4706]: I1208 19:54:40.608844 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:54:40 crc kubenswrapper[4706]: E1208 19:54:40.610128 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:54:49 crc kubenswrapper[4706]: I1208 19:54:49.051076 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6qpgg"] Dec 08 19:54:49 crc kubenswrapper[4706]: I1208 19:54:49.060581 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6qpgg"] Dec 08 19:54:49 crc kubenswrapper[4706]: I1208 19:54:49.625868 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8823e38-e8cd-47b1-8408-b999c691f189" path="/var/lib/kubelet/pods/e8823e38-e8cd-47b1-8408-b999c691f189/volumes" Dec 08 19:54:52 crc kubenswrapper[4706]: I1208 19:54:52.608480 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:54:52 crc kubenswrapper[4706]: E1208 19:54:52.609091 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 19:55:06 crc kubenswrapper[4706]: I1208 19:55:06.608383 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:55:06 crc kubenswrapper[4706]: I1208 19:55:06.995946 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"e06ecf241bbf1f6acfa682ccb5a43e38dd0a78858dbd6d48f37f0f9e8d629bc3"} Dec 08 19:55:11 crc kubenswrapper[4706]: I1208 19:55:11.587707 4706 scope.go:117] "RemoveContainer" containerID="092594639b173507d5ebac41d663c6f844dc9b6027306260eb017e99517ef3f8" Dec 08 19:55:11 crc kubenswrapper[4706]: I1208 19:55:11.630738 4706 scope.go:117] "RemoveContainer" containerID="b0244d3b41d9204a8c97ca3de1cb3259469af58b87513fd4711ccd12fffdfe62" Dec 08 19:55:11 crc kubenswrapper[4706]: I1208 19:55:11.675877 4706 scope.go:117] "RemoveContainer" containerID="1c626a309924b0d060b2ec6d4ee99df5f2fc5939f07d8b9caf0922aaf9202b18" Dec 08 19:55:11 crc kubenswrapper[4706]: I1208 19:55:11.723454 4706 scope.go:117] "RemoveContainer" containerID="b7900d848be946a04d35908bd11716c92feb0d454a870282e09895fe4d213f42" Dec 08 19:55:11 crc kubenswrapper[4706]: I1208 19:55:11.776027 4706 scope.go:117] "RemoveContainer" containerID="bd2970d9cf0ca99e8844cccb6d28536e32c753aae0789f12d236a6fec646977f" Dec 08 19:55:11 crc kubenswrapper[4706]: I1208 19:55:11.855637 4706 scope.go:117] "RemoveContainer" containerID="17832f12d8fd4aa04d54e49f062fac087c676ec1af1a1dbc40a930dc66c84c39" Dec 08 19:55:11 crc kubenswrapper[4706]: I1208 19:55:11.886306 4706 scope.go:117] "RemoveContainer" containerID="ce88419453a5ffff15eead9387f7089800790b5d2b71518b90c308d1d45a1a36" Dec 08 19:55:20 crc kubenswrapper[4706]: I1208 19:55:20.059504 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lkpfq"] Dec 08 19:55:20 crc kubenswrapper[4706]: I1208 19:55:20.070539 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lkpfq"] Dec 08 19:55:20 crc kubenswrapper[4706]: I1208 19:55:20.082467 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-5vj8x"] Dec 08 19:55:20 crc kubenswrapper[4706]: I1208 19:55:20.091459 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-5vj8x"] Dec 08 19:55:21 crc kubenswrapper[4706]: I1208 19:55:21.622938 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b2d5238-1fbe-4f49-b335-566d1ebaed31" path="/var/lib/kubelet/pods/3b2d5238-1fbe-4f49-b335-566d1ebaed31/volumes" Dec 08 19:55:21 crc kubenswrapper[4706]: I1208 19:55:21.624095 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab73a222-d68a-4716-b0bb-6750440338ad" path="/var/lib/kubelet/pods/ab73a222-d68a-4716-b0bb-6750440338ad/volumes" Dec 08 19:55:34 crc kubenswrapper[4706]: I1208 19:55:34.294369 4706 generic.go:334] "Generic (PLEG): container finished" podID="784135f7-ff54-43df-b7e7-ea4f0b940fdc" containerID="73e8ef0a119a93b618644622461dba2ef2d842900995883256490370eb44de6f" exitCode=0 Dec 08 19:55:34 crc kubenswrapper[4706]: I1208 19:55:34.294468 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" event={"ID":"784135f7-ff54-43df-b7e7-ea4f0b940fdc","Type":"ContainerDied","Data":"73e8ef0a119a93b618644622461dba2ef2d842900995883256490370eb44de6f"} Dec 08 19:55:35 crc kubenswrapper[4706]: I1208 19:55:35.940626 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.137119 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-inventory\") pod \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.137513 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bdrr\" (UniqueName: \"kubernetes.io/projected/784135f7-ff54-43df-b7e7-ea4f0b940fdc-kube-api-access-4bdrr\") pod \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.137539 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-ssh-key\") pod \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\" (UID: \"784135f7-ff54-43df-b7e7-ea4f0b940fdc\") " Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.145302 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/784135f7-ff54-43df-b7e7-ea4f0b940fdc-kube-api-access-4bdrr" (OuterVolumeSpecName: "kube-api-access-4bdrr") pod "784135f7-ff54-43df-b7e7-ea4f0b940fdc" (UID: "784135f7-ff54-43df-b7e7-ea4f0b940fdc"). InnerVolumeSpecName "kube-api-access-4bdrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.181088 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-inventory" (OuterVolumeSpecName: "inventory") pod "784135f7-ff54-43df-b7e7-ea4f0b940fdc" (UID: "784135f7-ff54-43df-b7e7-ea4f0b940fdc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.183768 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "784135f7-ff54-43df-b7e7-ea4f0b940fdc" (UID: "784135f7-ff54-43df-b7e7-ea4f0b940fdc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.239938 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bdrr\" (UniqueName: \"kubernetes.io/projected/784135f7-ff54-43df-b7e7-ea4f0b940fdc-kube-api-access-4bdrr\") on node \"crc\" DevicePath \"\"" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.239980 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.239993 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/784135f7-ff54-43df-b7e7-ea4f0b940fdc-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.317651 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" event={"ID":"784135f7-ff54-43df-b7e7-ea4f0b940fdc","Type":"ContainerDied","Data":"3b50c26b0d9565c688ebd8600a0508245875ed4af6d99b355c5c147c9e2b7382"} Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.317699 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b50c26b0d9565c688ebd8600a0508245875ed4af6d99b355c5c147c9e2b7382" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.317759 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.478378 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s"] Dec 08 19:55:36 crc kubenswrapper[4706]: E1208 19:55:36.479544 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="784135f7-ff54-43df-b7e7-ea4f0b940fdc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.479635 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="784135f7-ff54-43df-b7e7-ea4f0b940fdc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.479960 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="784135f7-ff54-43df-b7e7-ea4f0b940fdc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.481071 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.493038 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s"] Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.496802 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.496810 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.497070 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.497409 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.653610 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.654038 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4sz4\" (UniqueName: \"kubernetes.io/projected/7ea46b46-9aa7-46f3-a235-8eb100586d2f-kube-api-access-z4sz4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.654242 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.757279 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.757355 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4sz4\" (UniqueName: \"kubernetes.io/projected/7ea46b46-9aa7-46f3-a235-8eb100586d2f-kube-api-access-z4sz4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.757416 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.762285 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.776969 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.787158 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4sz4\" (UniqueName: \"kubernetes.io/projected/7ea46b46-9aa7-46f3-a235-8eb100586d2f-kube-api-access-z4sz4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:36 crc kubenswrapper[4706]: I1208 19:55:36.803204 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:37 crc kubenswrapper[4706]: I1208 19:55:37.423751 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s"] Dec 08 19:55:38 crc kubenswrapper[4706]: I1208 19:55:38.339932 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" event={"ID":"7ea46b46-9aa7-46f3-a235-8eb100586d2f","Type":"ContainerStarted","Data":"7b6335d36e0b1857840027b86f1ec27c23e627b81143757bd3598fa6119a2741"} Dec 08 19:55:38 crc kubenswrapper[4706]: I1208 19:55:38.340442 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" event={"ID":"7ea46b46-9aa7-46f3-a235-8eb100586d2f","Type":"ContainerStarted","Data":"7d2c51a623d5082849d2610968185cb3a38c348d2a86bb357722533a4af66c2f"} Dec 08 19:55:38 crc kubenswrapper[4706]: I1208 19:55:38.362202 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" podStartSLOduration=1.887982673 podStartE2EDuration="2.362179124s" podCreationTimestamp="2025-12-08 19:55:36 +0000 UTC" firstStartedPulling="2025-12-08 19:55:37.436824031 +0000 UTC m=+2020.079025034" lastFinishedPulling="2025-12-08 19:55:37.911020482 +0000 UTC m=+2020.553221485" observedRunningTime="2025-12-08 19:55:38.358478398 +0000 UTC m=+2021.000679411" watchObservedRunningTime="2025-12-08 19:55:38.362179124 +0000 UTC m=+2021.004380127" Dec 08 19:55:43 crc kubenswrapper[4706]: I1208 19:55:43.394589 4706 generic.go:334] "Generic (PLEG): container finished" podID="7ea46b46-9aa7-46f3-a235-8eb100586d2f" containerID="7b6335d36e0b1857840027b86f1ec27c23e627b81143757bd3598fa6119a2741" exitCode=0 Dec 08 19:55:43 crc kubenswrapper[4706]: I1208 19:55:43.394695 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" event={"ID":"7ea46b46-9aa7-46f3-a235-8eb100586d2f","Type":"ContainerDied","Data":"7b6335d36e0b1857840027b86f1ec27c23e627b81143757bd3598fa6119a2741"} Dec 08 19:55:44 crc kubenswrapper[4706]: I1208 19:55:44.901667 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:44 crc kubenswrapper[4706]: I1208 19:55:44.976111 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4sz4\" (UniqueName: \"kubernetes.io/projected/7ea46b46-9aa7-46f3-a235-8eb100586d2f-kube-api-access-z4sz4\") pod \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " Dec 08 19:55:44 crc kubenswrapper[4706]: I1208 19:55:44.976833 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-inventory\") pod \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " Dec 08 19:55:44 crc kubenswrapper[4706]: I1208 19:55:44.976896 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-ssh-key\") pod \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\" (UID: \"7ea46b46-9aa7-46f3-a235-8eb100586d2f\") " Dec 08 19:55:44 crc kubenswrapper[4706]: I1208 19:55:44.986582 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ea46b46-9aa7-46f3-a235-8eb100586d2f-kube-api-access-z4sz4" (OuterVolumeSpecName: "kube-api-access-z4sz4") pod "7ea46b46-9aa7-46f3-a235-8eb100586d2f" (UID: "7ea46b46-9aa7-46f3-a235-8eb100586d2f"). InnerVolumeSpecName "kube-api-access-z4sz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.048818 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7ea46b46-9aa7-46f3-a235-8eb100586d2f" (UID: "7ea46b46-9aa7-46f3-a235-8eb100586d2f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.075031 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-inventory" (OuterVolumeSpecName: "inventory") pod "7ea46b46-9aa7-46f3-a235-8eb100586d2f" (UID: "7ea46b46-9aa7-46f3-a235-8eb100586d2f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.090399 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4sz4\" (UniqueName: \"kubernetes.io/projected/7ea46b46-9aa7-46f3-a235-8eb100586d2f-kube-api-access-z4sz4\") on node \"crc\" DevicePath \"\"" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.090434 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.090442 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ea46b46-9aa7-46f3-a235-8eb100586d2f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.419159 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" event={"ID":"7ea46b46-9aa7-46f3-a235-8eb100586d2f","Type":"ContainerDied","Data":"7d2c51a623d5082849d2610968185cb3a38c348d2a86bb357722533a4af66c2f"} Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.419223 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d2c51a623d5082849d2610968185cb3a38c348d2a86bb357722533a4af66c2f" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.419346 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.525676 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26"] Dec 08 19:55:45 crc kubenswrapper[4706]: E1208 19:55:45.526693 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea46b46-9aa7-46f3-a235-8eb100586d2f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.526789 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea46b46-9aa7-46f3-a235-8eb100586d2f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.527216 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea46b46-9aa7-46f3-a235-8eb100586d2f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.528402 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.536282 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.536727 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.536893 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.536980 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26"] Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.537306 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.602009 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.602110 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.602144 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68hq9\" (UniqueName: \"kubernetes.io/projected/e30a59fd-9346-4fae-8a9d-14c8700fd608-kube-api-access-68hq9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.704169 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.704241 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68hq9\" (UniqueName: \"kubernetes.io/projected/e30a59fd-9346-4fae-8a9d-14c8700fd608-kube-api-access-68hq9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.704554 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.709677 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.710005 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.722345 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68hq9\" (UniqueName: \"kubernetes.io/projected/e30a59fd-9346-4fae-8a9d-14c8700fd608-kube-api-access-68hq9\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zjn26\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:45 crc kubenswrapper[4706]: I1208 19:55:45.853331 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:55:46 crc kubenswrapper[4706]: I1208 19:55:46.492533 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26"] Dec 08 19:55:47 crc kubenswrapper[4706]: I1208 19:55:47.449820 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" event={"ID":"e30a59fd-9346-4fae-8a9d-14c8700fd608","Type":"ContainerStarted","Data":"da479a8f8e0213cd09b89c2de221c2e90c78ba382a7993e1185b5e44b765a4aa"} Dec 08 19:55:47 crc kubenswrapper[4706]: I1208 19:55:47.450216 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" event={"ID":"e30a59fd-9346-4fae-8a9d-14c8700fd608","Type":"ContainerStarted","Data":"e03f5867896bad008c60f62000e1a786d43c2e78aa9d4a2c7eca8f1f585ddc37"} Dec 08 19:55:47 crc kubenswrapper[4706]: I1208 19:55:47.479723 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" podStartSLOduration=2.096448133 podStartE2EDuration="2.479703415s" podCreationTimestamp="2025-12-08 19:55:45 +0000 UTC" firstStartedPulling="2025-12-08 19:55:46.497339692 +0000 UTC m=+2029.139540695" lastFinishedPulling="2025-12-08 19:55:46.880594974 +0000 UTC m=+2029.522795977" observedRunningTime="2025-12-08 19:55:47.471276434 +0000 UTC m=+2030.113477447" watchObservedRunningTime="2025-12-08 19:55:47.479703415 +0000 UTC m=+2030.121904418" Dec 08 19:56:03 crc kubenswrapper[4706]: I1208 19:56:03.051293 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-lp5nt"] Dec 08 19:56:03 crc kubenswrapper[4706]: I1208 19:56:03.061982 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-lp5nt"] Dec 08 19:56:03 crc kubenswrapper[4706]: I1208 19:56:03.621994 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265" path="/var/lib/kubelet/pods/b706a41c-3dac-4ae7-b3fd-bf0d6e5ee265/volumes" Dec 08 19:56:12 crc kubenswrapper[4706]: I1208 19:56:12.087128 4706 scope.go:117] "RemoveContainer" containerID="9274343acb4c2eaaaf30b25d2d2ed7b107a3f84c9522cecba5d8e1e10b88fab5" Dec 08 19:56:12 crc kubenswrapper[4706]: I1208 19:56:12.158230 4706 scope.go:117] "RemoveContainer" containerID="8782ed0963e46f0dd0f6d8eb39ff25ef0919a61de7ff1869b910c1928c07a19a" Dec 08 19:56:12 crc kubenswrapper[4706]: I1208 19:56:12.224651 4706 scope.go:117] "RemoveContainer" containerID="47f68655d5e808bd4404524ff9da233aaf93866810ace1ed32992b1dd24e7c32" Dec 08 19:56:29 crc kubenswrapper[4706]: I1208 19:56:29.917291 4706 generic.go:334] "Generic (PLEG): container finished" podID="e30a59fd-9346-4fae-8a9d-14c8700fd608" containerID="da479a8f8e0213cd09b89c2de221c2e90c78ba382a7993e1185b5e44b765a4aa" exitCode=0 Dec 08 19:56:29 crc kubenswrapper[4706]: I1208 19:56:29.917406 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" event={"ID":"e30a59fd-9346-4fae-8a9d-14c8700fd608","Type":"ContainerDied","Data":"da479a8f8e0213cd09b89c2de221c2e90c78ba382a7993e1185b5e44b765a4aa"} Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.544415 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.641208 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-inventory\") pod \"e30a59fd-9346-4fae-8a9d-14c8700fd608\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.641437 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68hq9\" (UniqueName: \"kubernetes.io/projected/e30a59fd-9346-4fae-8a9d-14c8700fd608-kube-api-access-68hq9\") pod \"e30a59fd-9346-4fae-8a9d-14c8700fd608\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.641529 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-ssh-key\") pod \"e30a59fd-9346-4fae-8a9d-14c8700fd608\" (UID: \"e30a59fd-9346-4fae-8a9d-14c8700fd608\") " Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.668583 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e30a59fd-9346-4fae-8a9d-14c8700fd608-kube-api-access-68hq9" (OuterVolumeSpecName: "kube-api-access-68hq9") pod "e30a59fd-9346-4fae-8a9d-14c8700fd608" (UID: "e30a59fd-9346-4fae-8a9d-14c8700fd608"). InnerVolumeSpecName "kube-api-access-68hq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.734561 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-inventory" (OuterVolumeSpecName: "inventory") pod "e30a59fd-9346-4fae-8a9d-14c8700fd608" (UID: "e30a59fd-9346-4fae-8a9d-14c8700fd608"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.745417 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.745460 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68hq9\" (UniqueName: \"kubernetes.io/projected/e30a59fd-9346-4fae-8a9d-14c8700fd608-kube-api-access-68hq9\") on node \"crc\" DevicePath \"\"" Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.779545 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e30a59fd-9346-4fae-8a9d-14c8700fd608" (UID: "e30a59fd-9346-4fae-8a9d-14c8700fd608"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.855555 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e30a59fd-9346-4fae-8a9d-14c8700fd608-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.945858 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" event={"ID":"e30a59fd-9346-4fae-8a9d-14c8700fd608","Type":"ContainerDied","Data":"e03f5867896bad008c60f62000e1a786d43c2e78aa9d4a2c7eca8f1f585ddc37"} Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.946158 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zjn26" Dec 08 19:56:31 crc kubenswrapper[4706]: I1208 19:56:31.946291 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e03f5867896bad008c60f62000e1a786d43c2e78aa9d4a2c7eca8f1f585ddc37" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.065856 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh"] Dec 08 19:56:32 crc kubenswrapper[4706]: E1208 19:56:32.066481 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e30a59fd-9346-4fae-8a9d-14c8700fd608" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.066508 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e30a59fd-9346-4fae-8a9d-14c8700fd608" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.066809 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e30a59fd-9346-4fae-8a9d-14c8700fd608" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.067955 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.072198 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.072467 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.072592 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.079781 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh"] Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.080673 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.162909 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.163046 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.163217 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6fc9\" (UniqueName: \"kubernetes.io/projected/9576d118-0baa-462a-b75a-a31af5b84c66-kube-api-access-k6fc9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.265159 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6fc9\" (UniqueName: \"kubernetes.io/projected/9576d118-0baa-462a-b75a-a31af5b84c66-kube-api-access-k6fc9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.265348 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.266498 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.269634 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.270998 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.283080 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6fc9\" (UniqueName: \"kubernetes.io/projected/9576d118-0baa-462a-b75a-a31af5b84c66-kube-api-access-k6fc9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swwvh\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:32 crc kubenswrapper[4706]: I1208 19:56:32.409034 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:56:33 crc kubenswrapper[4706]: I1208 19:56:33.011149 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh"] Dec 08 19:56:33 crc kubenswrapper[4706]: I1208 19:56:33.971499 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" event={"ID":"9576d118-0baa-462a-b75a-a31af5b84c66","Type":"ContainerStarted","Data":"20ade4287960edb6a51f1d2bc488ee8ecb86b682f1b17c4b2523ea0a3fc38940"} Dec 08 19:56:33 crc kubenswrapper[4706]: I1208 19:56:33.972511 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" event={"ID":"9576d118-0baa-462a-b75a-a31af5b84c66","Type":"ContainerStarted","Data":"af02b6d7d1247b5ee6a95efa5bf82c63d6a1c24c4b31674f72edb5e88d028cce"} Dec 08 19:56:33 crc kubenswrapper[4706]: I1208 19:56:33.987338 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" podStartSLOduration=1.490599787 podStartE2EDuration="1.987316192s" podCreationTimestamp="2025-12-08 19:56:32 +0000 UTC" firstStartedPulling="2025-12-08 19:56:33.024007304 +0000 UTC m=+2075.666208307" lastFinishedPulling="2025-12-08 19:56:33.520723709 +0000 UTC m=+2076.162924712" observedRunningTime="2025-12-08 19:56:33.986658424 +0000 UTC m=+2076.628859427" watchObservedRunningTime="2025-12-08 19:56:33.987316192 +0000 UTC m=+2076.629517195" Dec 08 19:57:25 crc kubenswrapper[4706]: I1208 19:57:25.057989 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-wflqk"] Dec 08 19:57:25 crc kubenswrapper[4706]: I1208 19:57:25.072846 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-wflqk"] Dec 08 19:57:25 crc kubenswrapper[4706]: I1208 19:57:25.622793 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a92db262-e832-42a3-8339-4f1824075e10" path="/var/lib/kubelet/pods/a92db262-e832-42a3-8339-4f1824075e10/volumes" Dec 08 19:57:26 crc kubenswrapper[4706]: I1208 19:57:26.601774 4706 generic.go:334] "Generic (PLEG): container finished" podID="9576d118-0baa-462a-b75a-a31af5b84c66" containerID="20ade4287960edb6a51f1d2bc488ee8ecb86b682f1b17c4b2523ea0a3fc38940" exitCode=0 Dec 08 19:57:26 crc kubenswrapper[4706]: I1208 19:57:26.601821 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" event={"ID":"9576d118-0baa-462a-b75a-a31af5b84c66","Type":"ContainerDied","Data":"20ade4287960edb6a51f1d2bc488ee8ecb86b682f1b17c4b2523ea0a3fc38940"} Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.371143 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.459733 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-inventory\") pod \"9576d118-0baa-462a-b75a-a31af5b84c66\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.459839 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6fc9\" (UniqueName: \"kubernetes.io/projected/9576d118-0baa-462a-b75a-a31af5b84c66-kube-api-access-k6fc9\") pod \"9576d118-0baa-462a-b75a-a31af5b84c66\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.459882 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-ssh-key\") pod \"9576d118-0baa-462a-b75a-a31af5b84c66\" (UID: \"9576d118-0baa-462a-b75a-a31af5b84c66\") " Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.488683 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9576d118-0baa-462a-b75a-a31af5b84c66-kube-api-access-k6fc9" (OuterVolumeSpecName: "kube-api-access-k6fc9") pod "9576d118-0baa-462a-b75a-a31af5b84c66" (UID: "9576d118-0baa-462a-b75a-a31af5b84c66"). InnerVolumeSpecName "kube-api-access-k6fc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.520342 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9576d118-0baa-462a-b75a-a31af5b84c66" (UID: "9576d118-0baa-462a-b75a-a31af5b84c66"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.564214 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6fc9\" (UniqueName: \"kubernetes.io/projected/9576d118-0baa-462a-b75a-a31af5b84c66-kube-api-access-k6fc9\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.564275 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.590458 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-inventory" (OuterVolumeSpecName: "inventory") pod "9576d118-0baa-462a-b75a-a31af5b84c66" (UID: "9576d118-0baa-462a-b75a-a31af5b84c66"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.660746 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" event={"ID":"9576d118-0baa-462a-b75a-a31af5b84c66","Type":"ContainerDied","Data":"af02b6d7d1247b5ee6a95efa5bf82c63d6a1c24c4b31674f72edb5e88d028cce"} Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.660816 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af02b6d7d1247b5ee6a95efa5bf82c63d6a1c24c4b31674f72edb5e88d028cce" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.660909 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swwvh" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.665754 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9576d118-0baa-462a-b75a-a31af5b84c66-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.791874 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rhjj4"] Dec 08 19:57:28 crc kubenswrapper[4706]: E1208 19:57:28.792538 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9576d118-0baa-462a-b75a-a31af5b84c66" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.792565 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9576d118-0baa-462a-b75a-a31af5b84c66" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.792804 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9576d118-0baa-462a-b75a-a31af5b84c66" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.793712 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.808008 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.808031 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.808065 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.808010 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.810392 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rhjj4"] Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.870378 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.870511 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-255p5\" (UniqueName: \"kubernetes.io/projected/adf18455-0f5d-4d8e-a46d-c77523de9f84-kube-api-access-255p5\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.870550 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.973659 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.973832 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-255p5\" (UniqueName: \"kubernetes.io/projected/adf18455-0f5d-4d8e-a46d-c77523de9f84-kube-api-access-255p5\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.973884 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.980159 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.980638 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:28 crc kubenswrapper[4706]: I1208 19:57:28.994209 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-255p5\" (UniqueName: \"kubernetes.io/projected/adf18455-0f5d-4d8e-a46d-c77523de9f84-kube-api-access-255p5\") pod \"ssh-known-hosts-edpm-deployment-rhjj4\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:29 crc kubenswrapper[4706]: I1208 19:57:29.124827 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:29 crc kubenswrapper[4706]: I1208 19:57:29.792400 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rhjj4"] Dec 08 19:57:30 crc kubenswrapper[4706]: I1208 19:57:30.699607 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" event={"ID":"adf18455-0f5d-4d8e-a46d-c77523de9f84","Type":"ContainerStarted","Data":"2fdeb2dc88ea9fc55a1fe32b57fdfc2446d13a947cfbc6417eb08f30d89d4e8d"} Dec 08 19:57:30 crc kubenswrapper[4706]: I1208 19:57:30.699675 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" event={"ID":"adf18455-0f5d-4d8e-a46d-c77523de9f84","Type":"ContainerStarted","Data":"c7a859f3e038baf46dbdd1c7f930188d0f6d4a4b55fcc50fd175bb90e2775cf9"} Dec 08 19:57:30 crc kubenswrapper[4706]: I1208 19:57:30.721516 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" podStartSLOduration=2.329259437 podStartE2EDuration="2.721493483s" podCreationTimestamp="2025-12-08 19:57:28 +0000 UTC" firstStartedPulling="2025-12-08 19:57:29.79710276 +0000 UTC m=+2132.439303763" lastFinishedPulling="2025-12-08 19:57:30.189336806 +0000 UTC m=+2132.831537809" observedRunningTime="2025-12-08 19:57:30.720251028 +0000 UTC m=+2133.362452031" watchObservedRunningTime="2025-12-08 19:57:30.721493483 +0000 UTC m=+2133.363694486" Dec 08 19:57:31 crc kubenswrapper[4706]: I1208 19:57:31.037294 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-j2q44"] Dec 08 19:57:31 crc kubenswrapper[4706]: I1208 19:57:31.049067 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-j2q44"] Dec 08 19:57:31 crc kubenswrapper[4706]: I1208 19:57:31.660827 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98651db7-0458-47ab-8621-7098bf5337ae" path="/var/lib/kubelet/pods/98651db7-0458-47ab-8621-7098bf5337ae/volumes" Dec 08 19:57:35 crc kubenswrapper[4706]: I1208 19:57:35.835850 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:57:35 crc kubenswrapper[4706]: I1208 19:57:35.836719 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:57:37 crc kubenswrapper[4706]: I1208 19:57:37.780573 4706 generic.go:334] "Generic (PLEG): container finished" podID="adf18455-0f5d-4d8e-a46d-c77523de9f84" containerID="2fdeb2dc88ea9fc55a1fe32b57fdfc2446d13a947cfbc6417eb08f30d89d4e8d" exitCode=0 Dec 08 19:57:37 crc kubenswrapper[4706]: I1208 19:57:37.781021 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" event={"ID":"adf18455-0f5d-4d8e-a46d-c77523de9f84","Type":"ContainerDied","Data":"2fdeb2dc88ea9fc55a1fe32b57fdfc2446d13a947cfbc6417eb08f30d89d4e8d"} Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.401778 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.444813 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-inventory-0\") pod \"adf18455-0f5d-4d8e-a46d-c77523de9f84\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.445584 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-255p5\" (UniqueName: \"kubernetes.io/projected/adf18455-0f5d-4d8e-a46d-c77523de9f84-kube-api-access-255p5\") pod \"adf18455-0f5d-4d8e-a46d-c77523de9f84\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.445767 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-ssh-key-openstack-edpm-ipam\") pod \"adf18455-0f5d-4d8e-a46d-c77523de9f84\" (UID: \"adf18455-0f5d-4d8e-a46d-c77523de9f84\") " Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.461976 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf18455-0f5d-4d8e-a46d-c77523de9f84-kube-api-access-255p5" (OuterVolumeSpecName: "kube-api-access-255p5") pod "adf18455-0f5d-4d8e-a46d-c77523de9f84" (UID: "adf18455-0f5d-4d8e-a46d-c77523de9f84"). InnerVolumeSpecName "kube-api-access-255p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.484452 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "adf18455-0f5d-4d8e-a46d-c77523de9f84" (UID: "adf18455-0f5d-4d8e-a46d-c77523de9f84"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.496879 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "adf18455-0f5d-4d8e-a46d-c77523de9f84" (UID: "adf18455-0f5d-4d8e-a46d-c77523de9f84"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.549231 4706 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.549299 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-255p5\" (UniqueName: \"kubernetes.io/projected/adf18455-0f5d-4d8e-a46d-c77523de9f84-kube-api-access-255p5\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.549313 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf18455-0f5d-4d8e-a46d-c77523de9f84-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.803306 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" event={"ID":"adf18455-0f5d-4d8e-a46d-c77523de9f84","Type":"ContainerDied","Data":"c7a859f3e038baf46dbdd1c7f930188d0f6d4a4b55fcc50fd175bb90e2775cf9"} Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.803349 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rhjj4" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.803364 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a859f3e038baf46dbdd1c7f930188d0f6d4a4b55fcc50fd175bb90e2775cf9" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.926704 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5"] Dec 08 19:57:39 crc kubenswrapper[4706]: E1208 19:57:39.929414 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf18455-0f5d-4d8e-a46d-c77523de9f84" containerName="ssh-known-hosts-edpm-deployment" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.929450 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf18455-0f5d-4d8e-a46d-c77523de9f84" containerName="ssh-known-hosts-edpm-deployment" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.933724 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf18455-0f5d-4d8e-a46d-c77523de9f84" containerName="ssh-known-hosts-edpm-deployment" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.935036 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.946237 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.946398 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.946640 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.946783 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:57:39 crc kubenswrapper[4706]: I1208 19:57:39.950751 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5"] Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.062563 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9dfx\" (UniqueName: \"kubernetes.io/projected/893e5057-db47-44b6-b3da-f4e59a5992c5-kube-api-access-t9dfx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.062625 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.063099 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.165623 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9dfx\" (UniqueName: \"kubernetes.io/projected/893e5057-db47-44b6-b3da-f4e59a5992c5-kube-api-access-t9dfx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.165689 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.165797 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.170594 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.179141 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.193471 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9dfx\" (UniqueName: \"kubernetes.io/projected/893e5057-db47-44b6-b3da-f4e59a5992c5-kube-api-access-t9dfx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkcr5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.284130 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.409979 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5j84k"] Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.413675 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.447231 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5j84k"] Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.474947 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d4sd\" (UniqueName: \"kubernetes.io/projected/16ea442a-c4b9-4051-8252-21b3895886d4-kube-api-access-9d4sd\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.475040 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-utilities\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.475085 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-catalog-content\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.578370 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d4sd\" (UniqueName: \"kubernetes.io/projected/16ea442a-c4b9-4051-8252-21b3895886d4-kube-api-access-9d4sd\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.578468 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-utilities\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.578507 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-catalog-content\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.579250 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-catalog-content\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.579307 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-utilities\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.605929 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d4sd\" (UniqueName: \"kubernetes.io/projected/16ea442a-c4b9-4051-8252-21b3895886d4-kube-api-access-9d4sd\") pod \"certified-operators-5j84k\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.774025 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:40 crc kubenswrapper[4706]: I1208 19:57:40.996722 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5"] Dec 08 19:57:41 crc kubenswrapper[4706]: W1208 19:57:41.013086 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod893e5057_db47_44b6_b3da_f4e59a5992c5.slice/crio-30317fc6282d2e443e8e24c81cff41ca2f6b889568e4abf358fced15e40fdf01 WatchSource:0}: Error finding container 30317fc6282d2e443e8e24c81cff41ca2f6b889568e4abf358fced15e40fdf01: Status 404 returned error can't find the container with id 30317fc6282d2e443e8e24c81cff41ca2f6b889568e4abf358fced15e40fdf01 Dec 08 19:57:41 crc kubenswrapper[4706]: I1208 19:57:41.462067 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5j84k"] Dec 08 19:57:41 crc kubenswrapper[4706]: I1208 19:57:41.875604 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" event={"ID":"893e5057-db47-44b6-b3da-f4e59a5992c5","Type":"ContainerStarted","Data":"b33da0798b707c854ca747bd3823763a802714b3d037ed813669a5bbb8dd612a"} Dec 08 19:57:41 crc kubenswrapper[4706]: I1208 19:57:41.876096 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" event={"ID":"893e5057-db47-44b6-b3da-f4e59a5992c5","Type":"ContainerStarted","Data":"30317fc6282d2e443e8e24c81cff41ca2f6b889568e4abf358fced15e40fdf01"} Dec 08 19:57:41 crc kubenswrapper[4706]: I1208 19:57:41.879420 4706 generic.go:334] "Generic (PLEG): container finished" podID="16ea442a-c4b9-4051-8252-21b3895886d4" containerID="280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b" exitCode=0 Dec 08 19:57:41 crc kubenswrapper[4706]: I1208 19:57:41.879483 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j84k" event={"ID":"16ea442a-c4b9-4051-8252-21b3895886d4","Type":"ContainerDied","Data":"280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b"} Dec 08 19:57:41 crc kubenswrapper[4706]: I1208 19:57:41.879514 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j84k" event={"ID":"16ea442a-c4b9-4051-8252-21b3895886d4","Type":"ContainerStarted","Data":"a5fdf411c29272c125595dc4790a6b343e38ca24b9829dc19d5f99477f8328d9"} Dec 08 19:57:41 crc kubenswrapper[4706]: I1208 19:57:41.908331 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" podStartSLOduration=2.415755371 podStartE2EDuration="2.908300779s" podCreationTimestamp="2025-12-08 19:57:39 +0000 UTC" firstStartedPulling="2025-12-08 19:57:41.016279358 +0000 UTC m=+2143.658480351" lastFinishedPulling="2025-12-08 19:57:41.508824756 +0000 UTC m=+2144.151025759" observedRunningTime="2025-12-08 19:57:41.900334172 +0000 UTC m=+2144.542535185" watchObservedRunningTime="2025-12-08 19:57:41.908300779 +0000 UTC m=+2144.550501792" Dec 08 19:57:42 crc kubenswrapper[4706]: I1208 19:57:42.893290 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j84k" event={"ID":"16ea442a-c4b9-4051-8252-21b3895886d4","Type":"ContainerStarted","Data":"d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148"} Dec 08 19:57:44 crc kubenswrapper[4706]: I1208 19:57:44.915611 4706 generic.go:334] "Generic (PLEG): container finished" podID="16ea442a-c4b9-4051-8252-21b3895886d4" containerID="d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148" exitCode=0 Dec 08 19:57:44 crc kubenswrapper[4706]: I1208 19:57:44.915804 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j84k" event={"ID":"16ea442a-c4b9-4051-8252-21b3895886d4","Type":"ContainerDied","Data":"d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148"} Dec 08 19:57:45 crc kubenswrapper[4706]: I1208 19:57:45.933242 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j84k" event={"ID":"16ea442a-c4b9-4051-8252-21b3895886d4","Type":"ContainerStarted","Data":"5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf"} Dec 08 19:57:45 crc kubenswrapper[4706]: I1208 19:57:45.965705 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5j84k" podStartSLOduration=2.523995159 podStartE2EDuration="5.965674596s" podCreationTimestamp="2025-12-08 19:57:40 +0000 UTC" firstStartedPulling="2025-12-08 19:57:41.88202901 +0000 UTC m=+2144.524230013" lastFinishedPulling="2025-12-08 19:57:45.323708447 +0000 UTC m=+2147.965909450" observedRunningTime="2025-12-08 19:57:45.955559458 +0000 UTC m=+2148.597760471" watchObservedRunningTime="2025-12-08 19:57:45.965674596 +0000 UTC m=+2148.607875619" Dec 08 19:57:50 crc kubenswrapper[4706]: I1208 19:57:50.775001 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:50 crc kubenswrapper[4706]: I1208 19:57:50.776009 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:50 crc kubenswrapper[4706]: I1208 19:57:50.844225 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:50 crc kubenswrapper[4706]: I1208 19:57:50.987441 4706 generic.go:334] "Generic (PLEG): container finished" podID="893e5057-db47-44b6-b3da-f4e59a5992c5" containerID="b33da0798b707c854ca747bd3823763a802714b3d037ed813669a5bbb8dd612a" exitCode=0 Dec 08 19:57:50 crc kubenswrapper[4706]: I1208 19:57:50.987525 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" event={"ID":"893e5057-db47-44b6-b3da-f4e59a5992c5","Type":"ContainerDied","Data":"b33da0798b707c854ca747bd3823763a802714b3d037ed813669a5bbb8dd612a"} Dec 08 19:57:51 crc kubenswrapper[4706]: I1208 19:57:51.037093 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:51 crc kubenswrapper[4706]: I1208 19:57:51.095701 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5j84k"] Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.584959 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.632178 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-ssh-key\") pod \"893e5057-db47-44b6-b3da-f4e59a5992c5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.632924 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9dfx\" (UniqueName: \"kubernetes.io/projected/893e5057-db47-44b6-b3da-f4e59a5992c5-kube-api-access-t9dfx\") pod \"893e5057-db47-44b6-b3da-f4e59a5992c5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.633039 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-inventory\") pod \"893e5057-db47-44b6-b3da-f4e59a5992c5\" (UID: \"893e5057-db47-44b6-b3da-f4e59a5992c5\") " Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.649855 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/893e5057-db47-44b6-b3da-f4e59a5992c5-kube-api-access-t9dfx" (OuterVolumeSpecName: "kube-api-access-t9dfx") pod "893e5057-db47-44b6-b3da-f4e59a5992c5" (UID: "893e5057-db47-44b6-b3da-f4e59a5992c5"). InnerVolumeSpecName "kube-api-access-t9dfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.678241 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "893e5057-db47-44b6-b3da-f4e59a5992c5" (UID: "893e5057-db47-44b6-b3da-f4e59a5992c5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.684977 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-inventory" (OuterVolumeSpecName: "inventory") pod "893e5057-db47-44b6-b3da-f4e59a5992c5" (UID: "893e5057-db47-44b6-b3da-f4e59a5992c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.736369 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9dfx\" (UniqueName: \"kubernetes.io/projected/893e5057-db47-44b6-b3da-f4e59a5992c5-kube-api-access-t9dfx\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.736551 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:52 crc kubenswrapper[4706]: I1208 19:57:52.737193 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/893e5057-db47-44b6-b3da-f4e59a5992c5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.015498 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" event={"ID":"893e5057-db47-44b6-b3da-f4e59a5992c5","Type":"ContainerDied","Data":"30317fc6282d2e443e8e24c81cff41ca2f6b889568e4abf358fced15e40fdf01"} Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.015529 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkcr5" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.015566 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30317fc6282d2e443e8e24c81cff41ca2f6b889568e4abf358fced15e40fdf01" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.015735 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5j84k" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" containerName="registry-server" containerID="cri-o://5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf" gracePeriod=2 Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.115384 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f"] Dec 08 19:57:53 crc kubenswrapper[4706]: E1208 19:57:53.116121 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="893e5057-db47-44b6-b3da-f4e59a5992c5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.116146 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="893e5057-db47-44b6-b3da-f4e59a5992c5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.116525 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="893e5057-db47-44b6-b3da-f4e59a5992c5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.117893 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.120794 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.121203 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.121310 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.121522 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.127903 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f"] Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.147933 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vctfj\" (UniqueName: \"kubernetes.io/projected/5d9ea4d5-f428-449d-baf9-880c8a2637aa-kube-api-access-vctfj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.148436 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.148628 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.251048 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vctfj\" (UniqueName: \"kubernetes.io/projected/5d9ea4d5-f428-449d-baf9-880c8a2637aa-kube-api-access-vctfj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.251593 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.251681 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.259537 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.261025 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.269557 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vctfj\" (UniqueName: \"kubernetes.io/projected/5d9ea4d5-f428-449d-baf9-880c8a2637aa-kube-api-access-vctfj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.520778 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.651492 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.779882 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d4sd\" (UniqueName: \"kubernetes.io/projected/16ea442a-c4b9-4051-8252-21b3895886d4-kube-api-access-9d4sd\") pod \"16ea442a-c4b9-4051-8252-21b3895886d4\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.779973 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-utilities\") pod \"16ea442a-c4b9-4051-8252-21b3895886d4\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.779994 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-catalog-content\") pod \"16ea442a-c4b9-4051-8252-21b3895886d4\" (UID: \"16ea442a-c4b9-4051-8252-21b3895886d4\") " Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.781251 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-utilities" (OuterVolumeSpecName: "utilities") pod "16ea442a-c4b9-4051-8252-21b3895886d4" (UID: "16ea442a-c4b9-4051-8252-21b3895886d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.785975 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ea442a-c4b9-4051-8252-21b3895886d4-kube-api-access-9d4sd" (OuterVolumeSpecName: "kube-api-access-9d4sd") pod "16ea442a-c4b9-4051-8252-21b3895886d4" (UID: "16ea442a-c4b9-4051-8252-21b3895886d4"). InnerVolumeSpecName "kube-api-access-9d4sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.836581 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16ea442a-c4b9-4051-8252-21b3895886d4" (UID: "16ea442a-c4b9-4051-8252-21b3895886d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.882780 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d4sd\" (UniqueName: \"kubernetes.io/projected/16ea442a-c4b9-4051-8252-21b3895886d4-kube-api-access-9d4sd\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.882822 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:53 crc kubenswrapper[4706]: I1208 19:57:53.882832 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16ea442a-c4b9-4051-8252-21b3895886d4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.030104 4706 generic.go:334] "Generic (PLEG): container finished" podID="16ea442a-c4b9-4051-8252-21b3895886d4" containerID="5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf" exitCode=0 Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.030172 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j84k" event={"ID":"16ea442a-c4b9-4051-8252-21b3895886d4","Type":"ContainerDied","Data":"5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf"} Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.030204 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j84k" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.030279 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j84k" event={"ID":"16ea442a-c4b9-4051-8252-21b3895886d4","Type":"ContainerDied","Data":"a5fdf411c29272c125595dc4790a6b343e38ca24b9829dc19d5f99477f8328d9"} Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.030308 4706 scope.go:117] "RemoveContainer" containerID="5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.059005 4706 scope.go:117] "RemoveContainer" containerID="d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.083190 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5j84k"] Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.094473 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5j84k"] Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.113622 4706 scope.go:117] "RemoveContainer" containerID="280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.146474 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f"] Dec 08 19:57:54 crc kubenswrapper[4706]: W1208 19:57:54.152334 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d9ea4d5_f428_449d_baf9_880c8a2637aa.slice/crio-fe2e59fb850508d25095962d50983e1d4a13741b2df3331e97c52baf9218ef1c WatchSource:0}: Error finding container fe2e59fb850508d25095962d50983e1d4a13741b2df3331e97c52baf9218ef1c: Status 404 returned error can't find the container with id fe2e59fb850508d25095962d50983e1d4a13741b2df3331e97c52baf9218ef1c Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.163906 4706 scope.go:117] "RemoveContainer" containerID="5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf" Dec 08 19:57:54 crc kubenswrapper[4706]: E1208 19:57:54.164595 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf\": container with ID starting with 5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf not found: ID does not exist" containerID="5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.164641 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf"} err="failed to get container status \"5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf\": rpc error: code = NotFound desc = could not find container \"5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf\": container with ID starting with 5f83e73ecd3e4bae5ebaeb6d0e408aee87beb6c8a7bcac612183092d51d9bfbf not found: ID does not exist" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.164675 4706 scope.go:117] "RemoveContainer" containerID="d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148" Dec 08 19:57:54 crc kubenswrapper[4706]: E1208 19:57:54.165152 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148\": container with ID starting with d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148 not found: ID does not exist" containerID="d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.165213 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148"} err="failed to get container status \"d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148\": rpc error: code = NotFound desc = could not find container \"d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148\": container with ID starting with d86350f8338910db362702791602fa033256e799dc0aa7b9cbda2613d1a00148 not found: ID does not exist" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.165251 4706 scope.go:117] "RemoveContainer" containerID="280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b" Dec 08 19:57:54 crc kubenswrapper[4706]: E1208 19:57:54.165655 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b\": container with ID starting with 280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b not found: ID does not exist" containerID="280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b" Dec 08 19:57:54 crc kubenswrapper[4706]: I1208 19:57:54.165715 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b"} err="failed to get container status \"280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b\": rpc error: code = NotFound desc = could not find container \"280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b\": container with ID starting with 280bba265ca4936e857fbab9288028d74269f97340b807219b9abf359588609b not found: ID does not exist" Dec 08 19:57:55 crc kubenswrapper[4706]: I1208 19:57:55.049824 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" event={"ID":"5d9ea4d5-f428-449d-baf9-880c8a2637aa","Type":"ContainerStarted","Data":"e5ad5fe0928fe933a654dc65ea1f0b2f47ad2f4a057e2099f1f75f7daa7a292f"} Dec 08 19:57:55 crc kubenswrapper[4706]: I1208 19:57:55.050206 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" event={"ID":"5d9ea4d5-f428-449d-baf9-880c8a2637aa","Type":"ContainerStarted","Data":"fe2e59fb850508d25095962d50983e1d4a13741b2df3331e97c52baf9218ef1c"} Dec 08 19:57:55 crc kubenswrapper[4706]: I1208 19:57:55.084452 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" podStartSLOduration=1.681875671 podStartE2EDuration="2.084417241s" podCreationTimestamp="2025-12-08 19:57:53 +0000 UTC" firstStartedPulling="2025-12-08 19:57:54.171553367 +0000 UTC m=+2156.813754370" lastFinishedPulling="2025-12-08 19:57:54.574094937 +0000 UTC m=+2157.216295940" observedRunningTime="2025-12-08 19:57:55.07071557 +0000 UTC m=+2157.712916593" watchObservedRunningTime="2025-12-08 19:57:55.084417241 +0000 UTC m=+2157.726618254" Dec 08 19:57:55 crc kubenswrapper[4706]: I1208 19:57:55.624691 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" path="/var/lib/kubelet/pods/16ea442a-c4b9-4051-8252-21b3895886d4/volumes" Dec 08 19:58:05 crc kubenswrapper[4706]: I1208 19:58:05.160117 4706 generic.go:334] "Generic (PLEG): container finished" podID="5d9ea4d5-f428-449d-baf9-880c8a2637aa" containerID="e5ad5fe0928fe933a654dc65ea1f0b2f47ad2f4a057e2099f1f75f7daa7a292f" exitCode=0 Dec 08 19:58:05 crc kubenswrapper[4706]: I1208 19:58:05.160229 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" event={"ID":"5d9ea4d5-f428-449d-baf9-880c8a2637aa","Type":"ContainerDied","Data":"e5ad5fe0928fe933a654dc65ea1f0b2f47ad2f4a057e2099f1f75f7daa7a292f"} Dec 08 19:58:05 crc kubenswrapper[4706]: I1208 19:58:05.836010 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:58:05 crc kubenswrapper[4706]: I1208 19:58:05.836391 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.686575 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.817723 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-inventory\") pod \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.818017 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-ssh-key\") pod \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.818977 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vctfj\" (UniqueName: \"kubernetes.io/projected/5d9ea4d5-f428-449d-baf9-880c8a2637aa-kube-api-access-vctfj\") pod \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\" (UID: \"5d9ea4d5-f428-449d-baf9-880c8a2637aa\") " Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.828665 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d9ea4d5-f428-449d-baf9-880c8a2637aa-kube-api-access-vctfj" (OuterVolumeSpecName: "kube-api-access-vctfj") pod "5d9ea4d5-f428-449d-baf9-880c8a2637aa" (UID: "5d9ea4d5-f428-449d-baf9-880c8a2637aa"). InnerVolumeSpecName "kube-api-access-vctfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.857917 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d9ea4d5-f428-449d-baf9-880c8a2637aa" (UID: "5d9ea4d5-f428-449d-baf9-880c8a2637aa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.860274 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-inventory" (OuterVolumeSpecName: "inventory") pod "5d9ea4d5-f428-449d-baf9-880c8a2637aa" (UID: "5d9ea4d5-f428-449d-baf9-880c8a2637aa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.923107 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.923167 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vctfj\" (UniqueName: \"kubernetes.io/projected/5d9ea4d5-f428-449d-baf9-880c8a2637aa-kube-api-access-vctfj\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:06 crc kubenswrapper[4706]: I1208 19:58:06.923189 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d9ea4d5-f428-449d-baf9-880c8a2637aa-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.184692 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" event={"ID":"5d9ea4d5-f428-449d-baf9-880c8a2637aa","Type":"ContainerDied","Data":"fe2e59fb850508d25095962d50983e1d4a13741b2df3331e97c52baf9218ef1c"} Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.184760 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe2e59fb850508d25095962d50983e1d4a13741b2df3331e97c52baf9218ef1c" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.184766 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.363470 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4"] Dec 08 19:58:07 crc kubenswrapper[4706]: E1208 19:58:07.364069 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" containerName="extract-utilities" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.364095 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" containerName="extract-utilities" Dec 08 19:58:07 crc kubenswrapper[4706]: E1208 19:58:07.364114 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" containerName="extract-content" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.364125 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" containerName="extract-content" Dec 08 19:58:07 crc kubenswrapper[4706]: E1208 19:58:07.364145 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" containerName="registry-server" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.364155 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" containerName="registry-server" Dec 08 19:58:07 crc kubenswrapper[4706]: E1208 19:58:07.364202 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d9ea4d5-f428-449d-baf9-880c8a2637aa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.364211 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d9ea4d5-f428-449d-baf9-880c8a2637aa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.364470 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d9ea4d5-f428-449d-baf9-880c8a2637aa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.364501 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ea442a-c4b9-4051-8252-21b3895886d4" containerName="registry-server" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.365592 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.368688 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.368772 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.369008 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.369816 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.369824 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.369863 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.369889 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.369908 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.382661 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4"] Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.436617 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.437047 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.437181 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.437381 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.437601 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.437838 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fhmv\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-kube-api-access-5fhmv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.438166 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.438373 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.438419 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.438464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.438494 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.438681 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.438860 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.438988 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.541937 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542437 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542463 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542502 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542581 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542634 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fhmv\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-kube-api-access-5fhmv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542676 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542717 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542735 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542758 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542782 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542837 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542879 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.542926 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.549588 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.549600 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.549663 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.549792 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.549926 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.550281 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.550470 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.550733 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.551182 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.551216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.551407 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.551961 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.552339 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.566361 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fhmv\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-kube-api-access-5fhmv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:07 crc kubenswrapper[4706]: I1208 19:58:07.700843 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:08 crc kubenswrapper[4706]: I1208 19:58:08.313137 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4"] Dec 08 19:58:08 crc kubenswrapper[4706]: W1208 19:58:08.314464 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42663da5_6fa0_4332_8222_63fd50d7484e.slice/crio-639c8dfb048f89819c08a0c1e948449c552d2db83e86ccef0cffc5ddb7e64292 WatchSource:0}: Error finding container 639c8dfb048f89819c08a0c1e948449c552d2db83e86ccef0cffc5ddb7e64292: Status 404 returned error can't find the container with id 639c8dfb048f89819c08a0c1e948449c552d2db83e86ccef0cffc5ddb7e64292 Dec 08 19:58:09 crc kubenswrapper[4706]: I1208 19:58:09.211489 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" event={"ID":"42663da5-6fa0-4332-8222-63fd50d7484e","Type":"ContainerStarted","Data":"58012c692ce7dd0421cc4ee97f4dc147833e62ae4cc6e91d52e9e1f447f898a2"} Dec 08 19:58:09 crc kubenswrapper[4706]: I1208 19:58:09.212920 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" event={"ID":"42663da5-6fa0-4332-8222-63fd50d7484e","Type":"ContainerStarted","Data":"639c8dfb048f89819c08a0c1e948449c552d2db83e86ccef0cffc5ddb7e64292"} Dec 08 19:58:09 crc kubenswrapper[4706]: I1208 19:58:09.247148 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" podStartSLOduration=1.803499127 podStartE2EDuration="2.247114108s" podCreationTimestamp="2025-12-08 19:58:07 +0000 UTC" firstStartedPulling="2025-12-08 19:58:08.319165484 +0000 UTC m=+2170.961366487" lastFinishedPulling="2025-12-08 19:58:08.762780465 +0000 UTC m=+2171.404981468" observedRunningTime="2025-12-08 19:58:09.234788377 +0000 UTC m=+2171.876989380" watchObservedRunningTime="2025-12-08 19:58:09.247114108 +0000 UTC m=+2171.889315111" Dec 08 19:58:12 crc kubenswrapper[4706]: I1208 19:58:12.396503 4706 scope.go:117] "RemoveContainer" containerID="94bcd813bba47a429faf7a12e2270b23d2300b5b7e5e233de6ee69fc552c9097" Dec 08 19:58:12 crc kubenswrapper[4706]: I1208 19:58:12.442720 4706 scope.go:117] "RemoveContainer" containerID="e0ceaabd43655b7fa5356093f0759d3fbfe41f5fec8ec7421e65583bc11a0d44" Dec 08 19:58:14 crc kubenswrapper[4706]: I1208 19:58:14.925229 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8n8m7"] Dec 08 19:58:14 crc kubenswrapper[4706]: I1208 19:58:14.929518 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:14 crc kubenswrapper[4706]: I1208 19:58:14.958521 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8n8m7"] Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.052072 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-utilities\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.052899 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghn9\" (UniqueName: \"kubernetes.io/projected/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-kube-api-access-5ghn9\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.053115 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-catalog-content\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.155200 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-catalog-content\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.155617 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-utilities\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.155794 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghn9\" (UniqueName: \"kubernetes.io/projected/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-kube-api-access-5ghn9\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.155787 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-catalog-content\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.156235 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-utilities\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.180575 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghn9\" (UniqueName: \"kubernetes.io/projected/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-kube-api-access-5ghn9\") pod \"redhat-marketplace-8n8m7\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.255544 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:15 crc kubenswrapper[4706]: I1208 19:58:15.783929 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8n8m7"] Dec 08 19:58:16 crc kubenswrapper[4706]: I1208 19:58:16.309360 4706 generic.go:334] "Generic (PLEG): container finished" podID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerID="13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1" exitCode=0 Dec 08 19:58:16 crc kubenswrapper[4706]: I1208 19:58:16.309546 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n8m7" event={"ID":"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3","Type":"ContainerDied","Data":"13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1"} Dec 08 19:58:16 crc kubenswrapper[4706]: I1208 19:58:16.310014 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n8m7" event={"ID":"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3","Type":"ContainerStarted","Data":"bc11f557730e4c81fda8629429c595b980b112ddd13130df4e204692d76e8766"} Dec 08 19:58:17 crc kubenswrapper[4706]: I1208 19:58:17.321559 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n8m7" event={"ID":"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3","Type":"ContainerStarted","Data":"dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f"} Dec 08 19:58:18 crc kubenswrapper[4706]: I1208 19:58:18.335285 4706 generic.go:334] "Generic (PLEG): container finished" podID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerID="dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f" exitCode=0 Dec 08 19:58:18 crc kubenswrapper[4706]: I1208 19:58:18.335529 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n8m7" event={"ID":"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3","Type":"ContainerDied","Data":"dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f"} Dec 08 19:58:19 crc kubenswrapper[4706]: I1208 19:58:19.387199 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n8m7" event={"ID":"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3","Type":"ContainerStarted","Data":"ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0"} Dec 08 19:58:19 crc kubenswrapper[4706]: I1208 19:58:19.418143 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8n8m7" podStartSLOduration=2.9909651310000003 podStartE2EDuration="5.418103784s" podCreationTimestamp="2025-12-08 19:58:14 +0000 UTC" firstStartedPulling="2025-12-08 19:58:16.313420588 +0000 UTC m=+2178.955621591" lastFinishedPulling="2025-12-08 19:58:18.740559241 +0000 UTC m=+2181.382760244" observedRunningTime="2025-12-08 19:58:19.411821044 +0000 UTC m=+2182.054022067" watchObservedRunningTime="2025-12-08 19:58:19.418103784 +0000 UTC m=+2182.060304787" Dec 08 19:58:25 crc kubenswrapper[4706]: I1208 19:58:25.255768 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:25 crc kubenswrapper[4706]: I1208 19:58:25.256555 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:25 crc kubenswrapper[4706]: I1208 19:58:25.308737 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:25 crc kubenswrapper[4706]: I1208 19:58:25.538776 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:25 crc kubenswrapper[4706]: I1208 19:58:25.622648 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8n8m7"] Dec 08 19:58:27 crc kubenswrapper[4706]: I1208 19:58:27.503523 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8n8m7" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerName="registry-server" containerID="cri-o://ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0" gracePeriod=2 Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.067430 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.102145 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-catalog-content\") pod \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.102403 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ghn9\" (UniqueName: \"kubernetes.io/projected/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-kube-api-access-5ghn9\") pod \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.102502 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-utilities\") pod \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\" (UID: \"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3\") " Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.119253 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-utilities" (OuterVolumeSpecName: "utilities") pod "7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" (UID: "7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.142698 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-kube-api-access-5ghn9" (OuterVolumeSpecName: "kube-api-access-5ghn9") pod "7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" (UID: "7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3"). InnerVolumeSpecName "kube-api-access-5ghn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.153699 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" (UID: "7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.206295 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.206363 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ghn9\" (UniqueName: \"kubernetes.io/projected/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-kube-api-access-5ghn9\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.206379 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.517977 4706 generic.go:334] "Generic (PLEG): container finished" podID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerID="ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0" exitCode=0 Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.518044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n8m7" event={"ID":"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3","Type":"ContainerDied","Data":"ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0"} Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.518084 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n8m7" event={"ID":"7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3","Type":"ContainerDied","Data":"bc11f557730e4c81fda8629429c595b980b112ddd13130df4e204692d76e8766"} Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.518115 4706 scope.go:117] "RemoveContainer" containerID="ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.518159 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8n8m7" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.544745 4706 scope.go:117] "RemoveContainer" containerID="dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.563381 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8n8m7"] Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.572017 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8n8m7"] Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.589536 4706 scope.go:117] "RemoveContainer" containerID="13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.649128 4706 scope.go:117] "RemoveContainer" containerID="ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0" Dec 08 19:58:28 crc kubenswrapper[4706]: E1208 19:58:28.650071 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0\": container with ID starting with ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0 not found: ID does not exist" containerID="ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.650225 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0"} err="failed to get container status \"ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0\": rpc error: code = NotFound desc = could not find container \"ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0\": container with ID starting with ba257315e93ebf9b1f7e2e8db7a945b5bfe7cb1e4304682a8cc6c008d5ba31f0 not found: ID does not exist" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.650384 4706 scope.go:117] "RemoveContainer" containerID="dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f" Dec 08 19:58:28 crc kubenswrapper[4706]: E1208 19:58:28.651003 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f\": container with ID starting with dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f not found: ID does not exist" containerID="dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.651047 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f"} err="failed to get container status \"dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f\": rpc error: code = NotFound desc = could not find container \"dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f\": container with ID starting with dd077dc7e235883dfdd3821ce9bf46341a185f1dad054ccf3f4e7b18fd21793f not found: ID does not exist" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.651084 4706 scope.go:117] "RemoveContainer" containerID="13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1" Dec 08 19:58:28 crc kubenswrapper[4706]: E1208 19:58:28.651487 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1\": container with ID starting with 13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1 not found: ID does not exist" containerID="13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1" Dec 08 19:58:28 crc kubenswrapper[4706]: I1208 19:58:28.651598 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1"} err="failed to get container status \"13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1\": rpc error: code = NotFound desc = could not find container \"13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1\": container with ID starting with 13059a27780f029f79350b9d8c2816f31624f261b7b35c771d9415cc54fc08d1 not found: ID does not exist" Dec 08 19:58:29 crc kubenswrapper[4706]: I1208 19:58:29.621933 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" path="/var/lib/kubelet/pods/7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3/volumes" Dec 08 19:58:35 crc kubenswrapper[4706]: I1208 19:58:35.835860 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 19:58:35 crc kubenswrapper[4706]: I1208 19:58:35.836859 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 19:58:35 crc kubenswrapper[4706]: I1208 19:58:35.836933 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 19:58:35 crc kubenswrapper[4706]: I1208 19:58:35.838355 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e06ecf241bbf1f6acfa682ccb5a43e38dd0a78858dbd6d48f37f0f9e8d629bc3"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 19:58:35 crc kubenswrapper[4706]: I1208 19:58:35.838438 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://e06ecf241bbf1f6acfa682ccb5a43e38dd0a78858dbd6d48f37f0f9e8d629bc3" gracePeriod=600 Dec 08 19:58:36 crc kubenswrapper[4706]: I1208 19:58:36.609180 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="e06ecf241bbf1f6acfa682ccb5a43e38dd0a78858dbd6d48f37f0f9e8d629bc3" exitCode=0 Dec 08 19:58:36 crc kubenswrapper[4706]: I1208 19:58:36.609243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"e06ecf241bbf1f6acfa682ccb5a43e38dd0a78858dbd6d48f37f0f9e8d629bc3"} Dec 08 19:58:36 crc kubenswrapper[4706]: I1208 19:58:36.610051 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d"} Dec 08 19:58:36 crc kubenswrapper[4706]: I1208 19:58:36.610075 4706 scope.go:117] "RemoveContainer" containerID="29d15ecb0af7ba84f003bb58f1d6e69787cea231327d189bd6863cfe9b618494" Dec 08 19:58:50 crc kubenswrapper[4706]: I1208 19:58:50.791733 4706 generic.go:334] "Generic (PLEG): container finished" podID="42663da5-6fa0-4332-8222-63fd50d7484e" containerID="58012c692ce7dd0421cc4ee97f4dc147833e62ae4cc6e91d52e9e1f447f898a2" exitCode=0 Dec 08 19:58:50 crc kubenswrapper[4706]: I1208 19:58:50.792625 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" event={"ID":"42663da5-6fa0-4332-8222-63fd50d7484e","Type":"ContainerDied","Data":"58012c692ce7dd0421cc4ee97f4dc147833e62ae4cc6e91d52e9e1f447f898a2"} Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.396315 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.473446 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.473575 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.473641 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-nova-combined-ca-bundle\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.473680 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ssh-key\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.473777 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.473861 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ovn-combined-ca-bundle\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.473937 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-neutron-metadata-combined-ca-bundle\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.473961 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-libvirt-combined-ca-bundle\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.474018 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-repo-setup-combined-ca-bundle\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.474046 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fhmv\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-kube-api-access-5fhmv\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.474159 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-inventory\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.474257 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-telemetry-combined-ca-bundle\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.474345 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.474382 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-bootstrap-combined-ca-bundle\") pod \"42663da5-6fa0-4332-8222-63fd50d7484e\" (UID: \"42663da5-6fa0-4332-8222-63fd50d7484e\") " Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.486370 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.489230 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.492152 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.492186 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.493723 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.493626 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.503058 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.503109 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.503205 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.503297 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-kube-api-access-5fhmv" (OuterVolumeSpecName: "kube-api-access-5fhmv") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "kube-api-access-5fhmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.503327 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.504110 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.521396 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.535366 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-inventory" (OuterVolumeSpecName: "inventory") pod "42663da5-6fa0-4332-8222-63fd50d7484e" (UID: "42663da5-6fa0-4332-8222-63fd50d7484e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577137 4706 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577179 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fhmv\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-kube-api-access-5fhmv\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577192 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577206 4706 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577216 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577229 4706 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577239 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577250 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577273 4706 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577283 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577296 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/42663da5-6fa0-4332-8222-63fd50d7484e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577308 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577319 4706 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.577328 4706 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42663da5-6fa0-4332-8222-63fd50d7484e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.821482 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" event={"ID":"42663da5-6fa0-4332-8222-63fd50d7484e","Type":"ContainerDied","Data":"639c8dfb048f89819c08a0c1e948449c552d2db83e86ccef0cffc5ddb7e64292"} Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.821531 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="639c8dfb048f89819c08a0c1e948449c552d2db83e86ccef0cffc5ddb7e64292" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.821599 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.943302 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx"] Dec 08 19:58:52 crc kubenswrapper[4706]: E1208 19:58:52.943839 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerName="extract-content" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.943861 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerName="extract-content" Dec 08 19:58:52 crc kubenswrapper[4706]: E1208 19:58:52.943883 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerName="extract-utilities" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.943892 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerName="extract-utilities" Dec 08 19:58:52 crc kubenswrapper[4706]: E1208 19:58:52.943911 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerName="registry-server" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.943918 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerName="registry-server" Dec 08 19:58:52 crc kubenswrapper[4706]: E1208 19:58:52.943937 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42663da5-6fa0-4332-8222-63fd50d7484e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.943945 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="42663da5-6fa0-4332-8222-63fd50d7484e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.944168 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="42663da5-6fa0-4332-8222-63fd50d7484e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.944199 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce3b0ae-8865-4da7-a2a7-0e6d946dcaf3" containerName="registry-server" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.945076 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.947941 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.948170 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.949158 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.950364 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.961476 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 08 19:58:52 crc kubenswrapper[4706]: I1208 19:58:52.962652 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx"] Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.091408 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.091539 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.091616 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.091813 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.091965 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg7l9\" (UniqueName: \"kubernetes.io/projected/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-kube-api-access-xg7l9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.195761 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.195871 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.195924 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.195951 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.196009 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg7l9\" (UniqueName: \"kubernetes.io/projected/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-kube-api-access-xg7l9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.198282 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.202586 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.203391 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.206655 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.214181 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg7l9\" (UniqueName: \"kubernetes.io/projected/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-kube-api-access-xg7l9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zxsvx\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.265061 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 19:58:53 crc kubenswrapper[4706]: I1208 19:58:53.841318 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx"] Dec 08 19:58:54 crc kubenswrapper[4706]: I1208 19:58:54.848197 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" event={"ID":"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78","Type":"ContainerStarted","Data":"84d0d03d2999f087514c0f90e9104027439f96a2b1d9e56e72566f43678dff30"} Dec 08 19:58:54 crc kubenswrapper[4706]: I1208 19:58:54.849092 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" event={"ID":"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78","Type":"ContainerStarted","Data":"5eceb393847c9f78892f96353053735671f03d1bab98b6b66effc55c3efaa699"} Dec 08 19:58:54 crc kubenswrapper[4706]: I1208 19:58:54.882776 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" podStartSLOduration=2.35768456 podStartE2EDuration="2.882745004s" podCreationTimestamp="2025-12-08 19:58:52 +0000 UTC" firstStartedPulling="2025-12-08 19:58:53.849530706 +0000 UTC m=+2216.491731729" lastFinishedPulling="2025-12-08 19:58:54.37459118 +0000 UTC m=+2217.016792173" observedRunningTime="2025-12-08 19:58:54.872401029 +0000 UTC m=+2217.514602032" watchObservedRunningTime="2025-12-08 19:58:54.882745004 +0000 UTC m=+2217.524946007" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.164708 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420400-594th"] Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.167889 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.170945 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.170982 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.184323 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420400-594th"] Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.263095 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8whzd"] Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.267428 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.276428 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8whzd"] Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.359169 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-utilities\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.359379 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-config-volume\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.359401 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-secret-volume\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.359455 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mffnl\" (UniqueName: \"kubernetes.io/projected/5d28d58e-8648-4d6b-952b-87717a223236-kube-api-access-mffnl\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.359472 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-catalog-content\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.359505 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78qwk\" (UniqueName: \"kubernetes.io/projected/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-kube-api-access-78qwk\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.461462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-catalog-content\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.461517 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mffnl\" (UniqueName: \"kubernetes.io/projected/5d28d58e-8648-4d6b-952b-87717a223236-kube-api-access-mffnl\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.461555 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78qwk\" (UniqueName: \"kubernetes.io/projected/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-kube-api-access-78qwk\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.461604 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-utilities\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.461763 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-config-volume\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.461790 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-secret-volume\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.462136 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-catalog-content\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.462139 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-utilities\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.463084 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-config-volume\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.472325 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-secret-volume\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.486129 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78qwk\" (UniqueName: \"kubernetes.io/projected/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-kube-api-access-78qwk\") pod \"collect-profiles-29420400-594th\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.500600 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.500708 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mffnl\" (UniqueName: \"kubernetes.io/projected/5d28d58e-8648-4d6b-952b-87717a223236-kube-api-access-mffnl\") pod \"community-operators-8whzd\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.590695 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.738998 4706 generic.go:334] "Generic (PLEG): container finished" podID="2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" containerID="84d0d03d2999f087514c0f90e9104027439f96a2b1d9e56e72566f43678dff30" exitCode=0 Dec 08 20:00:00 crc kubenswrapper[4706]: I1208 20:00:00.739367 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" event={"ID":"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78","Type":"ContainerDied","Data":"84d0d03d2999f087514c0f90e9104027439f96a2b1d9e56e72566f43678dff30"} Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.144192 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8whzd"] Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.201962 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420400-594th"] Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.754000 4706 generic.go:334] "Generic (PLEG): container finished" podID="00924d56-a1ce-4bb5-a63f-fe296dc1bfad" containerID="0a387b9335d9e1995ced9ea80931994896933762a0b7652bbeb5ab51836c38d9" exitCode=0 Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.754089 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" event={"ID":"00924d56-a1ce-4bb5-a63f-fe296dc1bfad","Type":"ContainerDied","Data":"0a387b9335d9e1995ced9ea80931994896933762a0b7652bbeb5ab51836c38d9"} Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.754543 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" event={"ID":"00924d56-a1ce-4bb5-a63f-fe296dc1bfad","Type":"ContainerStarted","Data":"9556ca6836ff85e133314f15c15bd1ee2f9bf385c9fcd60e93ea5cbc932c9d2d"} Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.757598 4706 generic.go:334] "Generic (PLEG): container finished" podID="5d28d58e-8648-4d6b-952b-87717a223236" containerID="980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b" exitCode=0 Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.759087 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8whzd" event={"ID":"5d28d58e-8648-4d6b-952b-87717a223236","Type":"ContainerDied","Data":"980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b"} Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.759141 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8whzd" event={"ID":"5d28d58e-8648-4d6b-952b-87717a223236","Type":"ContainerStarted","Data":"e7cd6016cedbfdf103924555d867a300a2ac6e66c9356951639a2793db406e75"} Dec 08 20:00:01 crc kubenswrapper[4706]: I1208 20:00:01.761633 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.324474 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.515498 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovncontroller-config-0\") pod \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.515935 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovn-combined-ca-bundle\") pod \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.515976 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-inventory\") pod \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.516094 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ssh-key\") pod \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.516214 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg7l9\" (UniqueName: \"kubernetes.io/projected/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-kube-api-access-xg7l9\") pod \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\" (UID: \"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78\") " Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.529555 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" (UID: "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.529651 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-kube-api-access-xg7l9" (OuterVolumeSpecName: "kube-api-access-xg7l9") pod "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" (UID: "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78"). InnerVolumeSpecName "kube-api-access-xg7l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.560951 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" (UID: "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.570443 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-inventory" (OuterVolumeSpecName: "inventory") pod "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" (UID: "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.615493 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" (UID: "2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.620899 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.620996 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg7l9\" (UniqueName: \"kubernetes.io/projected/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-kube-api-access-xg7l9\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.621215 4706 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.621303 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.621365 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.773121 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" event={"ID":"2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78","Type":"ContainerDied","Data":"5eceb393847c9f78892f96353053735671f03d1bab98b6b66effc55c3efaa699"} Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.773219 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eceb393847c9f78892f96353053735671f03d1bab98b6b66effc55c3efaa699" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.773170 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zxsvx" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.869169 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb"] Dec 08 20:00:02 crc kubenswrapper[4706]: E1208 20:00:02.870217 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.870238 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.870518 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.871654 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.876514 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.876797 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.876968 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.877168 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.877364 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.877522 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.883671 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb"] Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.940653 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcg55\" (UniqueName: \"kubernetes.io/projected/543c86ef-f2f1-4b75-9de2-d613819b3e9a-kube-api-access-mcg55\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.940741 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.940876 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.940901 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.941132 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:02 crc kubenswrapper[4706]: I1208 20:00:02.941211 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.047223 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.047310 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.047374 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcg55\" (UniqueName: \"kubernetes.io/projected/543c86ef-f2f1-4b75-9de2-d613819b3e9a-kube-api-access-mcg55\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.047401 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.047453 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.047474 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.054526 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.055429 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.057544 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.058179 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.066054 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.067871 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcg55\" (UniqueName: \"kubernetes.io/projected/543c86ef-f2f1-4b75-9de2-d613819b3e9a-kube-api-access-mcg55\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.201565 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.201646 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.353898 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78qwk\" (UniqueName: \"kubernetes.io/projected/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-kube-api-access-78qwk\") pod \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.354064 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-config-volume\") pod \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.354255 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-secret-volume\") pod \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\" (UID: \"00924d56-a1ce-4bb5-a63f-fe296dc1bfad\") " Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.355129 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-config-volume" (OuterVolumeSpecName: "config-volume") pod "00924d56-a1ce-4bb5-a63f-fe296dc1bfad" (UID: "00924d56-a1ce-4bb5-a63f-fe296dc1bfad"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.358946 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "00924d56-a1ce-4bb5-a63f-fe296dc1bfad" (UID: "00924d56-a1ce-4bb5-a63f-fe296dc1bfad"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.359815 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.359853 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.361601 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-kube-api-access-78qwk" (OuterVolumeSpecName: "kube-api-access-78qwk") pod "00924d56-a1ce-4bb5-a63f-fe296dc1bfad" (UID: "00924d56-a1ce-4bb5-a63f-fe296dc1bfad"). InnerVolumeSpecName "kube-api-access-78qwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.462517 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78qwk\" (UniqueName: \"kubernetes.io/projected/00924d56-a1ce-4bb5-a63f-fe296dc1bfad-kube-api-access-78qwk\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.786461 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8whzd" event={"ID":"5d28d58e-8648-4d6b-952b-87717a223236","Type":"ContainerStarted","Data":"9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af"} Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.789286 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" event={"ID":"00924d56-a1ce-4bb5-a63f-fe296dc1bfad","Type":"ContainerDied","Data":"9556ca6836ff85e133314f15c15bd1ee2f9bf385c9fcd60e93ea5cbc932c9d2d"} Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.789347 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9556ca6836ff85e133314f15c15bd1ee2f9bf385c9fcd60e93ea5cbc932c9d2d" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.789414 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420400-594th" Dec 08 20:00:03 crc kubenswrapper[4706]: I1208 20:00:03.817631 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb"] Dec 08 20:00:03 crc kubenswrapper[4706]: W1208 20:00:03.828003 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod543c86ef_f2f1_4b75_9de2_d613819b3e9a.slice/crio-550791bcd8d5ccad0e4a81fe28b852b4ae8c6dd848886f1e911011a4f157343d WatchSource:0}: Error finding container 550791bcd8d5ccad0e4a81fe28b852b4ae8c6dd848886f1e911011a4f157343d: Status 404 returned error can't find the container with id 550791bcd8d5ccad0e4a81fe28b852b4ae8c6dd848886f1e911011a4f157343d Dec 08 20:00:04 crc kubenswrapper[4706]: I1208 20:00:04.304163 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r"] Dec 08 20:00:04 crc kubenswrapper[4706]: I1208 20:00:04.313634 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420355-cpd8r"] Dec 08 20:00:04 crc kubenswrapper[4706]: I1208 20:00:04.805986 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" event={"ID":"543c86ef-f2f1-4b75-9de2-d613819b3e9a","Type":"ContainerStarted","Data":"550791bcd8d5ccad0e4a81fe28b852b4ae8c6dd848886f1e911011a4f157343d"} Dec 08 20:00:04 crc kubenswrapper[4706]: I1208 20:00:04.828543 4706 generic.go:334] "Generic (PLEG): container finished" podID="5d28d58e-8648-4d6b-952b-87717a223236" containerID="9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af" exitCode=0 Dec 08 20:00:04 crc kubenswrapper[4706]: I1208 20:00:04.828603 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8whzd" event={"ID":"5d28d58e-8648-4d6b-952b-87717a223236","Type":"ContainerDied","Data":"9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af"} Dec 08 20:00:05 crc kubenswrapper[4706]: I1208 20:00:05.626517 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6889021e-b838-4fb3-b664-92efcdf699a2" path="/var/lib/kubelet/pods/6889021e-b838-4fb3-b664-92efcdf699a2/volumes" Dec 08 20:00:05 crc kubenswrapper[4706]: I1208 20:00:05.845021 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8whzd" event={"ID":"5d28d58e-8648-4d6b-952b-87717a223236","Type":"ContainerStarted","Data":"5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9"} Dec 08 20:00:05 crc kubenswrapper[4706]: I1208 20:00:05.849788 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" event={"ID":"543c86ef-f2f1-4b75-9de2-d613819b3e9a","Type":"ContainerStarted","Data":"1f6d135ad83bafc538f879d445373d238082221fb561ee2a047a6ccc4a9300d6"} Dec 08 20:00:05 crc kubenswrapper[4706]: I1208 20:00:05.875920 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8whzd" podStartSLOduration=2.326635652 podStartE2EDuration="5.875896323s" podCreationTimestamp="2025-12-08 20:00:00 +0000 UTC" firstStartedPulling="2025-12-08 20:00:01.761375333 +0000 UTC m=+2284.403576326" lastFinishedPulling="2025-12-08 20:00:05.310635994 +0000 UTC m=+2287.952836997" observedRunningTime="2025-12-08 20:00:05.866782224 +0000 UTC m=+2288.508983237" watchObservedRunningTime="2025-12-08 20:00:05.875896323 +0000 UTC m=+2288.518097326" Dec 08 20:00:05 crc kubenswrapper[4706]: I1208 20:00:05.888942 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" podStartSLOduration=3.27990385 podStartE2EDuration="3.888922914s" podCreationTimestamp="2025-12-08 20:00:02 +0000 UTC" firstStartedPulling="2025-12-08 20:00:03.831129964 +0000 UTC m=+2286.473330967" lastFinishedPulling="2025-12-08 20:00:04.440149028 +0000 UTC m=+2287.082350031" observedRunningTime="2025-12-08 20:00:05.886238698 +0000 UTC m=+2288.528439701" watchObservedRunningTime="2025-12-08 20:00:05.888922914 +0000 UTC m=+2288.531123917" Dec 08 20:00:10 crc kubenswrapper[4706]: I1208 20:00:10.591599 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:10 crc kubenswrapper[4706]: I1208 20:00:10.592551 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:10 crc kubenswrapper[4706]: I1208 20:00:10.649475 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:10 crc kubenswrapper[4706]: I1208 20:00:10.971013 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:11 crc kubenswrapper[4706]: I1208 20:00:11.034977 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8whzd"] Dec 08 20:00:12 crc kubenswrapper[4706]: I1208 20:00:12.629886 4706 scope.go:117] "RemoveContainer" containerID="3c30807b9e78027ef9f14262817bcc633e951ecc33add8e00cf7d96213ed582f" Dec 08 20:00:12 crc kubenswrapper[4706]: I1208 20:00:12.920018 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8whzd" podUID="5d28d58e-8648-4d6b-952b-87717a223236" containerName="registry-server" containerID="cri-o://5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9" gracePeriod=2 Dec 08 20:00:13 crc kubenswrapper[4706]: E1208 20:00:13.023684 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d28d58e_8648_4d6b_952b_87717a223236.slice/crio-5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9.scope\": RecentStats: unable to find data in memory cache]" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.595613 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.728383 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-utilities\") pod \"5d28d58e-8648-4d6b-952b-87717a223236\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.728620 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mffnl\" (UniqueName: \"kubernetes.io/projected/5d28d58e-8648-4d6b-952b-87717a223236-kube-api-access-mffnl\") pod \"5d28d58e-8648-4d6b-952b-87717a223236\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.728705 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-catalog-content\") pod \"5d28d58e-8648-4d6b-952b-87717a223236\" (UID: \"5d28d58e-8648-4d6b-952b-87717a223236\") " Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.730933 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-utilities" (OuterVolumeSpecName: "utilities") pod "5d28d58e-8648-4d6b-952b-87717a223236" (UID: "5d28d58e-8648-4d6b-952b-87717a223236"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.751673 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d28d58e-8648-4d6b-952b-87717a223236-kube-api-access-mffnl" (OuterVolumeSpecName: "kube-api-access-mffnl") pod "5d28d58e-8648-4d6b-952b-87717a223236" (UID: "5d28d58e-8648-4d6b-952b-87717a223236"). InnerVolumeSpecName "kube-api-access-mffnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.798654 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d28d58e-8648-4d6b-952b-87717a223236" (UID: "5d28d58e-8648-4d6b-952b-87717a223236"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.831465 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.831512 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d28d58e-8648-4d6b-952b-87717a223236-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.831523 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mffnl\" (UniqueName: \"kubernetes.io/projected/5d28d58e-8648-4d6b-952b-87717a223236-kube-api-access-mffnl\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.933624 4706 generic.go:334] "Generic (PLEG): container finished" podID="5d28d58e-8648-4d6b-952b-87717a223236" containerID="5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9" exitCode=0 Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.933671 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8whzd" event={"ID":"5d28d58e-8648-4d6b-952b-87717a223236","Type":"ContainerDied","Data":"5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9"} Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.933707 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8whzd" event={"ID":"5d28d58e-8648-4d6b-952b-87717a223236","Type":"ContainerDied","Data":"e7cd6016cedbfdf103924555d867a300a2ac6e66c9356951639a2793db406e75"} Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.933729 4706 scope.go:117] "RemoveContainer" containerID="5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.933810 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8whzd" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.963838 4706 scope.go:117] "RemoveContainer" containerID="9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af" Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.972371 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8whzd"] Dec 08 20:00:13 crc kubenswrapper[4706]: I1208 20:00:13.985170 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8whzd"] Dec 08 20:00:14 crc kubenswrapper[4706]: I1208 20:00:14.008739 4706 scope.go:117] "RemoveContainer" containerID="980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b" Dec 08 20:00:14 crc kubenswrapper[4706]: I1208 20:00:14.047029 4706 scope.go:117] "RemoveContainer" containerID="5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9" Dec 08 20:00:14 crc kubenswrapper[4706]: E1208 20:00:14.047716 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9\": container with ID starting with 5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9 not found: ID does not exist" containerID="5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9" Dec 08 20:00:14 crc kubenswrapper[4706]: I1208 20:00:14.047783 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9"} err="failed to get container status \"5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9\": rpc error: code = NotFound desc = could not find container \"5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9\": container with ID starting with 5b2db7a0d1a3374af6668fb6751f4f17774b1c258f8c5887166e0ca8ac91e0e9 not found: ID does not exist" Dec 08 20:00:14 crc kubenswrapper[4706]: I1208 20:00:14.047835 4706 scope.go:117] "RemoveContainer" containerID="9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af" Dec 08 20:00:14 crc kubenswrapper[4706]: E1208 20:00:14.048606 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af\": container with ID starting with 9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af not found: ID does not exist" containerID="9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af" Dec 08 20:00:14 crc kubenswrapper[4706]: I1208 20:00:14.048685 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af"} err="failed to get container status \"9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af\": rpc error: code = NotFound desc = could not find container \"9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af\": container with ID starting with 9ad3210aac975d495038c103b3b2797e83aa35148b94de892ec76cb03bbae9af not found: ID does not exist" Dec 08 20:00:14 crc kubenswrapper[4706]: I1208 20:00:14.048731 4706 scope.go:117] "RemoveContainer" containerID="980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b" Dec 08 20:00:14 crc kubenswrapper[4706]: E1208 20:00:14.049419 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b\": container with ID starting with 980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b not found: ID does not exist" containerID="980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b" Dec 08 20:00:14 crc kubenswrapper[4706]: I1208 20:00:14.049465 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b"} err="failed to get container status \"980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b\": rpc error: code = NotFound desc = could not find container \"980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b\": container with ID starting with 980b77b2954f23dd924f044b697b334f7dfe30a725e9eb96d81942dd935d816b not found: ID does not exist" Dec 08 20:00:15 crc kubenswrapper[4706]: I1208 20:00:15.622467 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d28d58e-8648-4d6b-952b-87717a223236" path="/var/lib/kubelet/pods/5d28d58e-8648-4d6b-952b-87717a223236/volumes" Dec 08 20:00:54 crc kubenswrapper[4706]: I1208 20:00:54.436105 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" event={"ID":"543c86ef-f2f1-4b75-9de2-d613819b3e9a","Type":"ContainerDied","Data":"1f6d135ad83bafc538f879d445373d238082221fb561ee2a047a6ccc4a9300d6"} Dec 08 20:00:54 crc kubenswrapper[4706]: I1208 20:00:54.436033 4706 generic.go:334] "Generic (PLEG): container finished" podID="543c86ef-f2f1-4b75-9de2-d613819b3e9a" containerID="1f6d135ad83bafc538f879d445373d238082221fb561ee2a047a6ccc4a9300d6" exitCode=0 Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.024697 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.117539 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-inventory\") pod \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.117699 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-nova-metadata-neutron-config-0\") pod \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.118503 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcg55\" (UniqueName: \"kubernetes.io/projected/543c86ef-f2f1-4b75-9de2-d613819b3e9a-kube-api-access-mcg55\") pod \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.118634 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-ssh-key\") pod \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.118777 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.118955 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-metadata-combined-ca-bundle\") pod \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\" (UID: \"543c86ef-f2f1-4b75-9de2-d613819b3e9a\") " Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.125279 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "543c86ef-f2f1-4b75-9de2-d613819b3e9a" (UID: "543c86ef-f2f1-4b75-9de2-d613819b3e9a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.125873 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/543c86ef-f2f1-4b75-9de2-d613819b3e9a-kube-api-access-mcg55" (OuterVolumeSpecName: "kube-api-access-mcg55") pod "543c86ef-f2f1-4b75-9de2-d613819b3e9a" (UID: "543c86ef-f2f1-4b75-9de2-d613819b3e9a"). InnerVolumeSpecName "kube-api-access-mcg55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.154576 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "543c86ef-f2f1-4b75-9de2-d613819b3e9a" (UID: "543c86ef-f2f1-4b75-9de2-d613819b3e9a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.164633 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-inventory" (OuterVolumeSpecName: "inventory") pod "543c86ef-f2f1-4b75-9de2-d613819b3e9a" (UID: "543c86ef-f2f1-4b75-9de2-d613819b3e9a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.166484 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "543c86ef-f2f1-4b75-9de2-d613819b3e9a" (UID: "543c86ef-f2f1-4b75-9de2-d613819b3e9a"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.181600 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "543c86ef-f2f1-4b75-9de2-d613819b3e9a" (UID: "543c86ef-f2f1-4b75-9de2-d613819b3e9a"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.222883 4706 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.222959 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcg55\" (UniqueName: \"kubernetes.io/projected/543c86ef-f2f1-4b75-9de2-d613819b3e9a-kube-api-access-mcg55\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.222979 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.222995 4706 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.223017 4706 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.223039 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/543c86ef-f2f1-4b75-9de2-d613819b3e9a-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.460740 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" event={"ID":"543c86ef-f2f1-4b75-9de2-d613819b3e9a","Type":"ContainerDied","Data":"550791bcd8d5ccad0e4a81fe28b852b4ae8c6dd848886f1e911011a4f157343d"} Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.460786 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="550791bcd8d5ccad0e4a81fe28b852b4ae8c6dd848886f1e911011a4f157343d" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.460826 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.663457 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn"] Dec 08 20:00:56 crc kubenswrapper[4706]: E1208 20:00:56.663966 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d28d58e-8648-4d6b-952b-87717a223236" containerName="registry-server" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.663989 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d28d58e-8648-4d6b-952b-87717a223236" containerName="registry-server" Dec 08 20:00:56 crc kubenswrapper[4706]: E1208 20:00:56.664009 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d28d58e-8648-4d6b-952b-87717a223236" containerName="extract-content" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.664015 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d28d58e-8648-4d6b-952b-87717a223236" containerName="extract-content" Dec 08 20:00:56 crc kubenswrapper[4706]: E1208 20:00:56.664037 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="543c86ef-f2f1-4b75-9de2-d613819b3e9a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.664044 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="543c86ef-f2f1-4b75-9de2-d613819b3e9a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 08 20:00:56 crc kubenswrapper[4706]: E1208 20:00:56.664058 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d28d58e-8648-4d6b-952b-87717a223236" containerName="extract-utilities" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.664064 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d28d58e-8648-4d6b-952b-87717a223236" containerName="extract-utilities" Dec 08 20:00:56 crc kubenswrapper[4706]: E1208 20:00:56.664079 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00924d56-a1ce-4bb5-a63f-fe296dc1bfad" containerName="collect-profiles" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.664085 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="00924d56-a1ce-4bb5-a63f-fe296dc1bfad" containerName="collect-profiles" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.664336 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d28d58e-8648-4d6b-952b-87717a223236" containerName="registry-server" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.664389 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="00924d56-a1ce-4bb5-a63f-fe296dc1bfad" containerName="collect-profiles" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.664405 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="543c86ef-f2f1-4b75-9de2-d613819b3e9a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.665281 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.667903 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.667903 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.668151 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.670227 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.677809 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.682507 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn"] Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.736500 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb5vm\" (UniqueName: \"kubernetes.io/projected/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-kube-api-access-fb5vm\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.736987 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.737236 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.737680 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.737862 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.842485 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.842616 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.842689 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb5vm\" (UniqueName: \"kubernetes.io/projected/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-kube-api-access-fb5vm\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.842735 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.842950 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.851005 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.851005 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.851256 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.851437 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.861316 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb5vm\" (UniqueName: \"kubernetes.io/projected/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-kube-api-access-fb5vm\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:56 crc kubenswrapper[4706]: I1208 20:00:56.985831 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:00:57 crc kubenswrapper[4706]: I1208 20:00:57.603734 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn"] Dec 08 20:00:58 crc kubenswrapper[4706]: I1208 20:00:58.062907 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 20:00:58 crc kubenswrapper[4706]: I1208 20:00:58.563862 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" event={"ID":"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8","Type":"ContainerStarted","Data":"d715f069803370df521e9d28dd5d4a74372b48320e1bc8822e9eb6c9a0a8da6f"} Dec 08 20:00:59 crc kubenswrapper[4706]: I1208 20:00:59.582154 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" event={"ID":"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8","Type":"ContainerStarted","Data":"319ab3486366e7a4e76863b1c7bcfd461d9f968cbc99bc53c48cc68e973d1bc9"} Dec 08 20:00:59 crc kubenswrapper[4706]: I1208 20:00:59.609922 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" podStartSLOduration=3.157230838 podStartE2EDuration="3.609893162s" podCreationTimestamp="2025-12-08 20:00:56 +0000 UTC" firstStartedPulling="2025-12-08 20:00:57.606457508 +0000 UTC m=+2340.248658511" lastFinishedPulling="2025-12-08 20:00:58.059119832 +0000 UTC m=+2340.701320835" observedRunningTime="2025-12-08 20:00:59.600616808 +0000 UTC m=+2342.242817811" watchObservedRunningTime="2025-12-08 20:00:59.609893162 +0000 UTC m=+2342.252094165" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.154242 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29420401-b95cm"] Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.156287 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.170928 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29420401-b95cm"] Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.252670 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-combined-ca-bundle\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.252897 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-config-data\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.253209 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjrj7\" (UniqueName: \"kubernetes.io/projected/68bc049a-f060-4de4-8be5-c88c55abebd9-kube-api-access-vjrj7\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.253377 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-fernet-keys\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.355678 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-fernet-keys\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.356292 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-combined-ca-bundle\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.356363 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-config-data\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.356502 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjrj7\" (UniqueName: \"kubernetes.io/projected/68bc049a-f060-4de4-8be5-c88c55abebd9-kube-api-access-vjrj7\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.364477 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-combined-ca-bundle\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.365570 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-config-data\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.367501 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-fernet-keys\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.382215 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjrj7\" (UniqueName: \"kubernetes.io/projected/68bc049a-f060-4de4-8be5-c88c55abebd9-kube-api-access-vjrj7\") pod \"keystone-cron-29420401-b95cm\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:00 crc kubenswrapper[4706]: I1208 20:01:00.483703 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:01 crc kubenswrapper[4706]: I1208 20:01:01.181609 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29420401-b95cm"] Dec 08 20:01:01 crc kubenswrapper[4706]: I1208 20:01:01.628147 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29420401-b95cm" event={"ID":"68bc049a-f060-4de4-8be5-c88c55abebd9","Type":"ContainerStarted","Data":"f8ba3966a3f6f8b7efcaa57d6bfd65936415cf3e9517ea52ff1957de5990a833"} Dec 08 20:01:01 crc kubenswrapper[4706]: I1208 20:01:01.628597 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29420401-b95cm" event={"ID":"68bc049a-f060-4de4-8be5-c88c55abebd9","Type":"ContainerStarted","Data":"972daec7e2cbe95bf143ad5e0987e4b102eb0f474289a7dddc9d85e8f90f76e3"} Dec 08 20:01:01 crc kubenswrapper[4706]: I1208 20:01:01.659746 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29420401-b95cm" podStartSLOduration=1.6597154650000001 podStartE2EDuration="1.659715465s" podCreationTimestamp="2025-12-08 20:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 20:01:01.644829771 +0000 UTC m=+2344.287030784" watchObservedRunningTime="2025-12-08 20:01:01.659715465 +0000 UTC m=+2344.301916468" Dec 08 20:01:04 crc kubenswrapper[4706]: E1208 20:01:04.564784 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68bc049a_f060_4de4_8be5_c88c55abebd9.slice/crio-f8ba3966a3f6f8b7efcaa57d6bfd65936415cf3e9517ea52ff1957de5990a833.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68bc049a_f060_4de4_8be5_c88c55abebd9.slice/crio-conmon-f8ba3966a3f6f8b7efcaa57d6bfd65936415cf3e9517ea52ff1957de5990a833.scope\": RecentStats: unable to find data in memory cache]" Dec 08 20:01:04 crc kubenswrapper[4706]: I1208 20:01:04.668564 4706 generic.go:334] "Generic (PLEG): container finished" podID="68bc049a-f060-4de4-8be5-c88c55abebd9" containerID="f8ba3966a3f6f8b7efcaa57d6bfd65936415cf3e9517ea52ff1957de5990a833" exitCode=0 Dec 08 20:01:04 crc kubenswrapper[4706]: I1208 20:01:04.668607 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29420401-b95cm" event={"ID":"68bc049a-f060-4de4-8be5-c88c55abebd9","Type":"ContainerDied","Data":"f8ba3966a3f6f8b7efcaa57d6bfd65936415cf3e9517ea52ff1957de5990a833"} Dec 08 20:01:05 crc kubenswrapper[4706]: I1208 20:01:05.836486 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:01:05 crc kubenswrapper[4706]: I1208 20:01:05.836956 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.250970 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.412634 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-combined-ca-bundle\") pod \"68bc049a-f060-4de4-8be5-c88c55abebd9\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.412879 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjrj7\" (UniqueName: \"kubernetes.io/projected/68bc049a-f060-4de4-8be5-c88c55abebd9-kube-api-access-vjrj7\") pod \"68bc049a-f060-4de4-8be5-c88c55abebd9\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.413045 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-config-data\") pod \"68bc049a-f060-4de4-8be5-c88c55abebd9\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.413117 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-fernet-keys\") pod \"68bc049a-f060-4de4-8be5-c88c55abebd9\" (UID: \"68bc049a-f060-4de4-8be5-c88c55abebd9\") " Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.428532 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68bc049a-f060-4de4-8be5-c88c55abebd9-kube-api-access-vjrj7" (OuterVolumeSpecName: "kube-api-access-vjrj7") pod "68bc049a-f060-4de4-8be5-c88c55abebd9" (UID: "68bc049a-f060-4de4-8be5-c88c55abebd9"). InnerVolumeSpecName "kube-api-access-vjrj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.428630 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "68bc049a-f060-4de4-8be5-c88c55abebd9" (UID: "68bc049a-f060-4de4-8be5-c88c55abebd9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.460297 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68bc049a-f060-4de4-8be5-c88c55abebd9" (UID: "68bc049a-f060-4de4-8be5-c88c55abebd9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.495178 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-config-data" (OuterVolumeSpecName: "config-data") pod "68bc049a-f060-4de4-8be5-c88c55abebd9" (UID: "68bc049a-f060-4de4-8be5-c88c55abebd9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.518363 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjrj7\" (UniqueName: \"kubernetes.io/projected/68bc049a-f060-4de4-8be5-c88c55abebd9-kube-api-access-vjrj7\") on node \"crc\" DevicePath \"\"" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.518735 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.518802 4706 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.518856 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68bc049a-f060-4de4-8be5-c88c55abebd9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.694743 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29420401-b95cm" event={"ID":"68bc049a-f060-4de4-8be5-c88c55abebd9","Type":"ContainerDied","Data":"972daec7e2cbe95bf143ad5e0987e4b102eb0f474289a7dddc9d85e8f90f76e3"} Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.694796 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="972daec7e2cbe95bf143ad5e0987e4b102eb0f474289a7dddc9d85e8f90f76e3" Dec 08 20:01:06 crc kubenswrapper[4706]: I1208 20:01:06.694836 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29420401-b95cm" Dec 08 20:01:35 crc kubenswrapper[4706]: I1208 20:01:35.835927 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:01:35 crc kubenswrapper[4706]: I1208 20:01:35.836987 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.014182 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kmvgb"] Dec 08 20:01:40 crc kubenswrapper[4706]: E1208 20:01:40.015799 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68bc049a-f060-4de4-8be5-c88c55abebd9" containerName="keystone-cron" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.015818 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="68bc049a-f060-4de4-8be5-c88c55abebd9" containerName="keystone-cron" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.016066 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="68bc049a-f060-4de4-8be5-c88c55abebd9" containerName="keystone-cron" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.017788 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.031655 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kmvgb"] Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.105645 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-catalog-content\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.105718 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-utilities\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.105970 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5ljc\" (UniqueName: \"kubernetes.io/projected/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-kube-api-access-b5ljc\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.208489 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-catalog-content\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.208568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-utilities\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.208701 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5ljc\" (UniqueName: \"kubernetes.io/projected/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-kube-api-access-b5ljc\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.209109 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-catalog-content\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.209614 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-utilities\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.242565 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5ljc\" (UniqueName: \"kubernetes.io/projected/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-kube-api-access-b5ljc\") pod \"redhat-operators-kmvgb\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.352116 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:40 crc kubenswrapper[4706]: I1208 20:01:40.889662 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kmvgb"] Dec 08 20:01:41 crc kubenswrapper[4706]: I1208 20:01:41.154776 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kmvgb" event={"ID":"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619","Type":"ContainerStarted","Data":"c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3"} Dec 08 20:01:41 crc kubenswrapper[4706]: I1208 20:01:41.156445 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kmvgb" event={"ID":"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619","Type":"ContainerStarted","Data":"6743fc7d8e03dc56128a4dc7285b9fbd5150f52ec03b0e786d591a9394e8c082"} Dec 08 20:01:42 crc kubenswrapper[4706]: I1208 20:01:42.167570 4706 generic.go:334] "Generic (PLEG): container finished" podID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerID="c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3" exitCode=0 Dec 08 20:01:42 crc kubenswrapper[4706]: I1208 20:01:42.167706 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kmvgb" event={"ID":"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619","Type":"ContainerDied","Data":"c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3"} Dec 08 20:01:44 crc kubenswrapper[4706]: I1208 20:01:44.207177 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kmvgb" event={"ID":"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619","Type":"ContainerStarted","Data":"5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9"} Dec 08 20:01:46 crc kubenswrapper[4706]: I1208 20:01:46.233529 4706 generic.go:334] "Generic (PLEG): container finished" podID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerID="5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9" exitCode=0 Dec 08 20:01:46 crc kubenswrapper[4706]: I1208 20:01:46.233602 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kmvgb" event={"ID":"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619","Type":"ContainerDied","Data":"5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9"} Dec 08 20:01:47 crc kubenswrapper[4706]: I1208 20:01:47.250847 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kmvgb" event={"ID":"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619","Type":"ContainerStarted","Data":"59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4"} Dec 08 20:01:47 crc kubenswrapper[4706]: I1208 20:01:47.273091 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kmvgb" podStartSLOduration=3.598100255 podStartE2EDuration="8.273072668s" podCreationTimestamp="2025-12-08 20:01:39 +0000 UTC" firstStartedPulling="2025-12-08 20:01:42.171256856 +0000 UTC m=+2384.813457869" lastFinishedPulling="2025-12-08 20:01:46.846229279 +0000 UTC m=+2389.488430282" observedRunningTime="2025-12-08 20:01:47.26964442 +0000 UTC m=+2389.911845443" watchObservedRunningTime="2025-12-08 20:01:47.273072668 +0000 UTC m=+2389.915273671" Dec 08 20:01:50 crc kubenswrapper[4706]: I1208 20:01:50.352907 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:50 crc kubenswrapper[4706]: I1208 20:01:50.353948 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:01:51 crc kubenswrapper[4706]: I1208 20:01:51.442927 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kmvgb" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="registry-server" probeResult="failure" output=< Dec 08 20:01:51 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 20:01:51 crc kubenswrapper[4706]: > Dec 08 20:02:00 crc kubenswrapper[4706]: I1208 20:02:00.414172 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:02:00 crc kubenswrapper[4706]: I1208 20:02:00.484369 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:02:00 crc kubenswrapper[4706]: I1208 20:02:00.664626 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kmvgb"] Dec 08 20:02:02 crc kubenswrapper[4706]: I1208 20:02:02.450737 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kmvgb" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="registry-server" containerID="cri-o://59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4" gracePeriod=2 Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.172944 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.329176 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5ljc\" (UniqueName: \"kubernetes.io/projected/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-kube-api-access-b5ljc\") pod \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.329410 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-catalog-content\") pod \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.329464 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-utilities\") pod \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\" (UID: \"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619\") " Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.330750 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-utilities" (OuterVolumeSpecName: "utilities") pod "8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" (UID: "8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.345596 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-kube-api-access-b5ljc" (OuterVolumeSpecName: "kube-api-access-b5ljc") pod "8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" (UID: "8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619"). InnerVolumeSpecName "kube-api-access-b5ljc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.432994 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5ljc\" (UniqueName: \"kubernetes.io/projected/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-kube-api-access-b5ljc\") on node \"crc\" DevicePath \"\"" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.433033 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.454376 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" (UID: "8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.469014 4706 generic.go:334] "Generic (PLEG): container finished" podID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerID="59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4" exitCode=0 Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.469091 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kmvgb" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.469092 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kmvgb" event={"ID":"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619","Type":"ContainerDied","Data":"59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4"} Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.469273 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kmvgb" event={"ID":"8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619","Type":"ContainerDied","Data":"6743fc7d8e03dc56128a4dc7285b9fbd5150f52ec03b0e786d591a9394e8c082"} Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.469304 4706 scope.go:117] "RemoveContainer" containerID="59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.517079 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kmvgb"] Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.526776 4706 scope.go:117] "RemoveContainer" containerID="5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.530685 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kmvgb"] Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.535583 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.559250 4706 scope.go:117] "RemoveContainer" containerID="c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.611913 4706 scope.go:117] "RemoveContainer" containerID="59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4" Dec 08 20:02:03 crc kubenswrapper[4706]: E1208 20:02:03.612811 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4\": container with ID starting with 59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4 not found: ID does not exist" containerID="59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.612849 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4"} err="failed to get container status \"59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4\": rpc error: code = NotFound desc = could not find container \"59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4\": container with ID starting with 59e132ff667bf042861d3173eda4efd51e74f1c4c9fd0ddfc31a1fb3eda91bf4 not found: ID does not exist" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.612874 4706 scope.go:117] "RemoveContainer" containerID="5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9" Dec 08 20:02:03 crc kubenswrapper[4706]: E1208 20:02:03.613111 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9\": container with ID starting with 5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9 not found: ID does not exist" containerID="5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.613137 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9"} err="failed to get container status \"5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9\": rpc error: code = NotFound desc = could not find container \"5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9\": container with ID starting with 5682e97845259dcfab81b7040110d5ab077b16bf6a5de3a203c266f83db592a9 not found: ID does not exist" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.613150 4706 scope.go:117] "RemoveContainer" containerID="c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3" Dec 08 20:02:03 crc kubenswrapper[4706]: E1208 20:02:03.613366 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3\": container with ID starting with c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3 not found: ID does not exist" containerID="c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.613393 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3"} err="failed to get container status \"c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3\": rpc error: code = NotFound desc = could not find container \"c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3\": container with ID starting with c7a233d3139b4b7b6122d811649992a3147beb437b7bd406036f3de1161ce5d3 not found: ID does not exist" Dec 08 20:02:03 crc kubenswrapper[4706]: I1208 20:02:03.623586 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" path="/var/lib/kubelet/pods/8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619/volumes" Dec 08 20:02:05 crc kubenswrapper[4706]: I1208 20:02:05.836492 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:02:05 crc kubenswrapper[4706]: I1208 20:02:05.837178 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:02:05 crc kubenswrapper[4706]: I1208 20:02:05.837247 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 20:02:05 crc kubenswrapper[4706]: I1208 20:02:05.838446 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 20:02:05 crc kubenswrapper[4706]: I1208 20:02:05.838525 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" gracePeriod=600 Dec 08 20:02:05 crc kubenswrapper[4706]: E1208 20:02:05.965936 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:02:06 crc kubenswrapper[4706]: I1208 20:02:06.502914 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" exitCode=0 Dec 08 20:02:06 crc kubenswrapper[4706]: I1208 20:02:06.502999 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d"} Dec 08 20:02:06 crc kubenswrapper[4706]: I1208 20:02:06.503093 4706 scope.go:117] "RemoveContainer" containerID="e06ecf241bbf1f6acfa682ccb5a43e38dd0a78858dbd6d48f37f0f9e8d629bc3" Dec 08 20:02:06 crc kubenswrapper[4706]: I1208 20:02:06.504055 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:02:06 crc kubenswrapper[4706]: E1208 20:02:06.504545 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:02:21 crc kubenswrapper[4706]: I1208 20:02:21.608994 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:02:21 crc kubenswrapper[4706]: E1208 20:02:21.610145 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:02:36 crc kubenswrapper[4706]: I1208 20:02:36.609651 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:02:36 crc kubenswrapper[4706]: E1208 20:02:36.610571 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:02:47 crc kubenswrapper[4706]: I1208 20:02:47.623750 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:02:47 crc kubenswrapper[4706]: E1208 20:02:47.625080 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:03:01 crc kubenswrapper[4706]: I1208 20:03:01.608555 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:03:01 crc kubenswrapper[4706]: E1208 20:03:01.609743 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:03:15 crc kubenswrapper[4706]: I1208 20:03:15.608996 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:03:15 crc kubenswrapper[4706]: E1208 20:03:15.610240 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:03:30 crc kubenswrapper[4706]: I1208 20:03:30.608237 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:03:30 crc kubenswrapper[4706]: E1208 20:03:30.609318 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:03:45 crc kubenswrapper[4706]: I1208 20:03:45.609114 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:03:45 crc kubenswrapper[4706]: E1208 20:03:45.610072 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:03:57 crc kubenswrapper[4706]: I1208 20:03:57.620170 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:03:57 crc kubenswrapper[4706]: E1208 20:03:57.621444 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:04:11 crc kubenswrapper[4706]: I1208 20:04:11.610901 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:04:11 crc kubenswrapper[4706]: E1208 20:04:11.611973 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:04:25 crc kubenswrapper[4706]: I1208 20:04:25.609044 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:04:25 crc kubenswrapper[4706]: E1208 20:04:25.610108 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:04:38 crc kubenswrapper[4706]: I1208 20:04:38.609080 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:04:38 crc kubenswrapper[4706]: E1208 20:04:38.609903 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:04:51 crc kubenswrapper[4706]: I1208 20:04:51.609400 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:04:51 crc kubenswrapper[4706]: E1208 20:04:51.610772 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:05:06 crc kubenswrapper[4706]: I1208 20:05:06.608725 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:05:06 crc kubenswrapper[4706]: E1208 20:05:06.611193 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:05:14 crc kubenswrapper[4706]: I1208 20:05:14.943582 4706 generic.go:334] "Generic (PLEG): container finished" podID="93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" containerID="319ab3486366e7a4e76863b1c7bcfd461d9f968cbc99bc53c48cc68e973d1bc9" exitCode=0 Dec 08 20:05:14 crc kubenswrapper[4706]: I1208 20:05:14.943717 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" event={"ID":"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8","Type":"ContainerDied","Data":"319ab3486366e7a4e76863b1c7bcfd461d9f968cbc99bc53c48cc68e973d1bc9"} Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.501006 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.551934 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-ssh-key\") pod \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.551990 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-combined-ca-bundle\") pod \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.552122 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-inventory\") pod \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.552149 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb5vm\" (UniqueName: \"kubernetes.io/projected/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-kube-api-access-fb5vm\") pod \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.552180 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-secret-0\") pod \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\" (UID: \"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8\") " Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.570873 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-kube-api-access-fb5vm" (OuterVolumeSpecName: "kube-api-access-fb5vm") pod "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" (UID: "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8"). InnerVolumeSpecName "kube-api-access-fb5vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.575575 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" (UID: "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.631410 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-inventory" (OuterVolumeSpecName: "inventory") pod "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" (UID: "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.654459 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" (UID: "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.670737 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.670786 4706 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.670805 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.670819 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb5vm\" (UniqueName: \"kubernetes.io/projected/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-kube-api-access-fb5vm\") on node \"crc\" DevicePath \"\"" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.672073 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" (UID: "93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.774483 4706 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.969175 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" event={"ID":"93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8","Type":"ContainerDied","Data":"d715f069803370df521e9d28dd5d4a74372b48320e1bc8822e9eb6c9a0a8da6f"} Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.969222 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d715f069803370df521e9d28dd5d4a74372b48320e1bc8822e9eb6c9a0a8da6f" Dec 08 20:05:16 crc kubenswrapper[4706]: I1208 20:05:16.969248 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.095638 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl"] Dec 08 20:05:17 crc kubenswrapper[4706]: E1208 20:05:17.098011 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.098048 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 08 20:05:17 crc kubenswrapper[4706]: E1208 20:05:17.098100 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="extract-content" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.098109 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="extract-content" Dec 08 20:05:17 crc kubenswrapper[4706]: E1208 20:05:17.098136 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="registry-server" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.098144 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="registry-server" Dec 08 20:05:17 crc kubenswrapper[4706]: E1208 20:05:17.098172 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="extract-utilities" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.098180 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="extract-utilities" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.099478 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bd1f3f7-6bf1-4aab-a87e-6f5b9f10d619" containerName="registry-server" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.099516 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.102296 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.106520 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.110753 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.111612 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.112017 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.112389 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.112929 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.113184 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.157491 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl"] Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.198765 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.198945 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.199075 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.199211 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.199591 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.199755 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.199906 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd4md\" (UniqueName: \"kubernetes.io/projected/43b41201-b854-4025-b70b-4455fcd8fc3c-kube-api-access-pd4md\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.200142 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.200399 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.302757 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.302854 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.302975 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.303025 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.303051 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.303090 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd4md\" (UniqueName: \"kubernetes.io/projected/43b41201-b854-4025-b70b-4455fcd8fc3c-kube-api-access-pd4md\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.303152 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.303200 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.303251 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.304059 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.307950 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.308000 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.308299 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.309209 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.309284 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.309501 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.311027 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.329178 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd4md\" (UniqueName: \"kubernetes.io/projected/43b41201-b854-4025-b70b-4455fcd8fc3c-kube-api-access-pd4md\") pod \"nova-edpm-deployment-openstack-edpm-ipam-m2jvl\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:17 crc kubenswrapper[4706]: I1208 20:05:17.468497 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:05:18 crc kubenswrapper[4706]: I1208 20:05:18.074440 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl"] Dec 08 20:05:18 crc kubenswrapper[4706]: I1208 20:05:18.080161 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 20:05:18 crc kubenswrapper[4706]: I1208 20:05:18.993170 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" event={"ID":"43b41201-b854-4025-b70b-4455fcd8fc3c","Type":"ContainerStarted","Data":"ba55a5b1d34a54074745f6598b4c1aceb254d57a311a285d2336b6c7ec69fd90"} Dec 08 20:05:18 crc kubenswrapper[4706]: I1208 20:05:18.993681 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" event={"ID":"43b41201-b854-4025-b70b-4455fcd8fc3c","Type":"ContainerStarted","Data":"d0ba7933c1b55d193af89f937f38a02ae2d152ed878450c710bc4ee8e84c58fb"} Dec 08 20:05:19 crc kubenswrapper[4706]: I1208 20:05:19.020166 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" podStartSLOduration=1.467612252 podStartE2EDuration="2.020140824s" podCreationTimestamp="2025-12-08 20:05:17 +0000 UTC" firstStartedPulling="2025-12-08 20:05:18.079819251 +0000 UTC m=+2600.722020254" lastFinishedPulling="2025-12-08 20:05:18.632347813 +0000 UTC m=+2601.274548826" observedRunningTime="2025-12-08 20:05:19.011398516 +0000 UTC m=+2601.653599529" watchObservedRunningTime="2025-12-08 20:05:19.020140824 +0000 UTC m=+2601.662341827" Dec 08 20:05:21 crc kubenswrapper[4706]: I1208 20:05:21.608405 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:05:21 crc kubenswrapper[4706]: E1208 20:05:21.609483 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:05:34 crc kubenswrapper[4706]: I1208 20:05:34.609360 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:05:34 crc kubenswrapper[4706]: E1208 20:05:34.610414 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:05:46 crc kubenswrapper[4706]: I1208 20:05:46.608550 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:05:46 crc kubenswrapper[4706]: E1208 20:05:46.609423 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:06:01 crc kubenswrapper[4706]: I1208 20:06:01.612547 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:06:01 crc kubenswrapper[4706]: E1208 20:06:01.613483 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:06:13 crc kubenswrapper[4706]: I1208 20:06:13.608980 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:06:13 crc kubenswrapper[4706]: E1208 20:06:13.610045 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:06:24 crc kubenswrapper[4706]: I1208 20:06:24.609743 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:06:24 crc kubenswrapper[4706]: E1208 20:06:24.610746 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:06:35 crc kubenswrapper[4706]: I1208 20:06:35.609121 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:06:35 crc kubenswrapper[4706]: E1208 20:06:35.610116 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:06:49 crc kubenswrapper[4706]: I1208 20:06:49.610944 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:06:49 crc kubenswrapper[4706]: E1208 20:06:49.613088 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:07:04 crc kubenswrapper[4706]: I1208 20:07:04.609398 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:07:04 crc kubenswrapper[4706]: E1208 20:07:04.610658 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:07:17 crc kubenswrapper[4706]: I1208 20:07:17.616931 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:07:18 crc kubenswrapper[4706]: I1208 20:07:18.383516 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"56b0d08effe4a9c86e0110c858d054b693a2be2056b7c9c944051a267d32813e"} Dec 08 20:08:15 crc kubenswrapper[4706]: I1208 20:08:15.072326 4706 generic.go:334] "Generic (PLEG): container finished" podID="43b41201-b854-4025-b70b-4455fcd8fc3c" containerID="ba55a5b1d34a54074745f6598b4c1aceb254d57a311a285d2336b6c7ec69fd90" exitCode=0 Dec 08 20:08:15 crc kubenswrapper[4706]: I1208 20:08:15.072422 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" event={"ID":"43b41201-b854-4025-b70b-4455fcd8fc3c","Type":"ContainerDied","Data":"ba55a5b1d34a54074745f6598b4c1aceb254d57a311a285d2336b6c7ec69fd90"} Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.684840 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.735895 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-0\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.736558 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-combined-ca-bundle\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.736776 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-ssh-key\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.737000 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-inventory\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.737222 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-1\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.737485 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd4md\" (UniqueName: \"kubernetes.io/projected/43b41201-b854-4025-b70b-4455fcd8fc3c-kube-api-access-pd4md\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.737614 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-1\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.738422 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-extra-config-0\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.738599 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-0\") pod \"43b41201-b854-4025-b70b-4455fcd8fc3c\" (UID: \"43b41201-b854-4025-b70b-4455fcd8fc3c\") " Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.747024 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43b41201-b854-4025-b70b-4455fcd8fc3c-kube-api-access-pd4md" (OuterVolumeSpecName: "kube-api-access-pd4md") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "kube-api-access-pd4md". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.749618 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.777526 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.786974 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.793748 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.803600 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.810486 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.821744 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-inventory" (OuterVolumeSpecName: "inventory") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.830863 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "43b41201-b854-4025-b70b-4455fcd8fc3c" (UID: "43b41201-b854-4025-b70b-4455fcd8fc3c"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848633 4706 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848674 4706 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848686 4706 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848695 4706 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848705 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848716 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848725 4706 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848736 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd4md\" (UniqueName: \"kubernetes.io/projected/43b41201-b854-4025-b70b-4455fcd8fc3c-kube-api-access-pd4md\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:16 crc kubenswrapper[4706]: I1208 20:08:16.848745 4706 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43b41201-b854-4025-b70b-4455fcd8fc3c-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.101603 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" event={"ID":"43b41201-b854-4025-b70b-4455fcd8fc3c","Type":"ContainerDied","Data":"d0ba7933c1b55d193af89f937f38a02ae2d152ed878450c710bc4ee8e84c58fb"} Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.101931 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0ba7933c1b55d193af89f937f38a02ae2d152ed878450c710bc4ee8e84c58fb" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.101708 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-m2jvl" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.223676 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p"] Dec 08 20:08:17 crc kubenswrapper[4706]: E1208 20:08:17.224538 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b41201-b854-4025-b70b-4455fcd8fc3c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.224616 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b41201-b854-4025-b70b-4455fcd8fc3c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.224916 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="43b41201-b854-4025-b70b-4455fcd8fc3c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.225785 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.231426 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.232217 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.233010 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8zhv" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.233297 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.233648 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.243836 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p"] Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.259660 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nbtt\" (UniqueName: \"kubernetes.io/projected/39dc4f79-e40f-479c-b82b-8ac18056c9ec-kube-api-access-5nbtt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.259743 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.259777 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.259820 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.259846 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.259925 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.259944 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.362551 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.362613 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.362706 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nbtt\" (UniqueName: \"kubernetes.io/projected/39dc4f79-e40f-479c-b82b-8ac18056c9ec-kube-api-access-5nbtt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.362774 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.362814 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.362862 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.362894 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.368185 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.368479 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.368744 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.369594 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.369793 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.370184 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.383641 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nbtt\" (UniqueName: \"kubernetes.io/projected/39dc4f79-e40f-479c-b82b-8ac18056c9ec-kube-api-access-5nbtt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:17 crc kubenswrapper[4706]: I1208 20:08:17.557048 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:08:18 crc kubenswrapper[4706]: I1208 20:08:18.205986 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p"] Dec 08 20:08:19 crc kubenswrapper[4706]: I1208 20:08:19.133517 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" event={"ID":"39dc4f79-e40f-479c-b82b-8ac18056c9ec","Type":"ContainerStarted","Data":"a32a13e0fbc8bf9d38fe5b9b49c1aabedccb63648a6baefff71d4ff1ed058dfc"} Dec 08 20:08:20 crc kubenswrapper[4706]: I1208 20:08:20.146115 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" event={"ID":"39dc4f79-e40f-479c-b82b-8ac18056c9ec","Type":"ContainerStarted","Data":"620febf1a23182cb8e1c72918678b48cd2a244073703e847e6f96fe169b6fd1e"} Dec 08 20:08:20 crc kubenswrapper[4706]: I1208 20:08:20.173884 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" podStartSLOduration=2.50039483 podStartE2EDuration="3.173862857s" podCreationTimestamp="2025-12-08 20:08:17 +0000 UTC" firstStartedPulling="2025-12-08 20:08:18.212427329 +0000 UTC m=+2780.854628332" lastFinishedPulling="2025-12-08 20:08:18.885895366 +0000 UTC m=+2781.528096359" observedRunningTime="2025-12-08 20:08:20.164564983 +0000 UTC m=+2782.806766056" watchObservedRunningTime="2025-12-08 20:08:20.173862857 +0000 UTC m=+2782.816063860" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.191322 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kzbqp"] Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.194671 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.211094 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kzbqp"] Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.242714 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-catalog-content\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.242919 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm22t\" (UniqueName: \"kubernetes.io/projected/a360f605-0014-46ee-973c-48d792d7e895-kube-api-access-dm22t\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.243302 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-utilities\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.346451 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-utilities\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.346651 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-catalog-content\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.346739 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm22t\" (UniqueName: \"kubernetes.io/projected/a360f605-0014-46ee-973c-48d792d7e895-kube-api-access-dm22t\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.347076 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-utilities\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.347549 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-catalog-content\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.376751 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm22t\" (UniqueName: \"kubernetes.io/projected/a360f605-0014-46ee-973c-48d792d7e895-kube-api-access-dm22t\") pod \"certified-operators-kzbqp\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:50 crc kubenswrapper[4706]: I1208 20:08:50.519070 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:08:51 crc kubenswrapper[4706]: I1208 20:08:51.140060 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kzbqp"] Dec 08 20:08:51 crc kubenswrapper[4706]: I1208 20:08:51.492034 4706 generic.go:334] "Generic (PLEG): container finished" podID="a360f605-0014-46ee-973c-48d792d7e895" containerID="28c2fe18dcf2704258987dd130093c4dd66e3b297b1deca1052b996bda5f2cbd" exitCode=0 Dec 08 20:08:51 crc kubenswrapper[4706]: I1208 20:08:51.492110 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzbqp" event={"ID":"a360f605-0014-46ee-973c-48d792d7e895","Type":"ContainerDied","Data":"28c2fe18dcf2704258987dd130093c4dd66e3b297b1deca1052b996bda5f2cbd"} Dec 08 20:08:51 crc kubenswrapper[4706]: I1208 20:08:51.493502 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzbqp" event={"ID":"a360f605-0014-46ee-973c-48d792d7e895","Type":"ContainerStarted","Data":"1678909b84eb0f5d8f256902020459fa4786ac5d8f5fc572b075540963d228f4"} Dec 08 20:08:52 crc kubenswrapper[4706]: I1208 20:08:52.507676 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzbqp" event={"ID":"a360f605-0014-46ee-973c-48d792d7e895","Type":"ContainerStarted","Data":"3090c530d69bc6e68720904dd8e0344b23546076e43458ad6254aaf6c8c48b52"} Dec 08 20:08:53 crc kubenswrapper[4706]: I1208 20:08:53.523740 4706 generic.go:334] "Generic (PLEG): container finished" podID="a360f605-0014-46ee-973c-48d792d7e895" containerID="3090c530d69bc6e68720904dd8e0344b23546076e43458ad6254aaf6c8c48b52" exitCode=0 Dec 08 20:08:53 crc kubenswrapper[4706]: I1208 20:08:53.523839 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzbqp" event={"ID":"a360f605-0014-46ee-973c-48d792d7e895","Type":"ContainerDied","Data":"3090c530d69bc6e68720904dd8e0344b23546076e43458ad6254aaf6c8c48b52"} Dec 08 20:08:54 crc kubenswrapper[4706]: I1208 20:08:54.537755 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzbqp" event={"ID":"a360f605-0014-46ee-973c-48d792d7e895","Type":"ContainerStarted","Data":"7f30bbcd9be30b2edfdb6933f5f463e5023cd57a2c77ad078dab76914e216176"} Dec 08 20:08:54 crc kubenswrapper[4706]: I1208 20:08:54.559916 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kzbqp" podStartSLOduration=2.106255372 podStartE2EDuration="4.559893844s" podCreationTimestamp="2025-12-08 20:08:50 +0000 UTC" firstStartedPulling="2025-12-08 20:08:51.495760851 +0000 UTC m=+2814.137961854" lastFinishedPulling="2025-12-08 20:08:53.949399323 +0000 UTC m=+2816.591600326" observedRunningTime="2025-12-08 20:08:54.557411554 +0000 UTC m=+2817.199612557" watchObservedRunningTime="2025-12-08 20:08:54.559893844 +0000 UTC m=+2817.202094847" Dec 08 20:09:00 crc kubenswrapper[4706]: I1208 20:09:00.519218 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:09:00 crc kubenswrapper[4706]: I1208 20:09:00.519889 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:09:00 crc kubenswrapper[4706]: I1208 20:09:00.575121 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:09:00 crc kubenswrapper[4706]: I1208 20:09:00.655539 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:09:00 crc kubenswrapper[4706]: I1208 20:09:00.826903 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kzbqp"] Dec 08 20:09:02 crc kubenswrapper[4706]: I1208 20:09:02.623963 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kzbqp" podUID="a360f605-0014-46ee-973c-48d792d7e895" containerName="registry-server" containerID="cri-o://7f30bbcd9be30b2edfdb6933f5f463e5023cd57a2c77ad078dab76914e216176" gracePeriod=2 Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.638248 4706 generic.go:334] "Generic (PLEG): container finished" podID="a360f605-0014-46ee-973c-48d792d7e895" containerID="7f30bbcd9be30b2edfdb6933f5f463e5023cd57a2c77ad078dab76914e216176" exitCode=0 Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.638305 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzbqp" event={"ID":"a360f605-0014-46ee-973c-48d792d7e895","Type":"ContainerDied","Data":"7f30bbcd9be30b2edfdb6933f5f463e5023cd57a2c77ad078dab76914e216176"} Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.639070 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzbqp" event={"ID":"a360f605-0014-46ee-973c-48d792d7e895","Type":"ContainerDied","Data":"1678909b84eb0f5d8f256902020459fa4786ac5d8f5fc572b075540963d228f4"} Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.639092 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1678909b84eb0f5d8f256902020459fa4786ac5d8f5fc572b075540963d228f4" Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.705924 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.827892 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-utilities\") pod \"a360f605-0014-46ee-973c-48d792d7e895\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.828031 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-catalog-content\") pod \"a360f605-0014-46ee-973c-48d792d7e895\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.828295 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm22t\" (UniqueName: \"kubernetes.io/projected/a360f605-0014-46ee-973c-48d792d7e895-kube-api-access-dm22t\") pod \"a360f605-0014-46ee-973c-48d792d7e895\" (UID: \"a360f605-0014-46ee-973c-48d792d7e895\") " Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.829315 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-utilities" (OuterVolumeSpecName: "utilities") pod "a360f605-0014-46ee-973c-48d792d7e895" (UID: "a360f605-0014-46ee-973c-48d792d7e895"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.836514 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a360f605-0014-46ee-973c-48d792d7e895-kube-api-access-dm22t" (OuterVolumeSpecName: "kube-api-access-dm22t") pod "a360f605-0014-46ee-973c-48d792d7e895" (UID: "a360f605-0014-46ee-973c-48d792d7e895"). InnerVolumeSpecName "kube-api-access-dm22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.882426 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a360f605-0014-46ee-973c-48d792d7e895" (UID: "a360f605-0014-46ee-973c-48d792d7e895"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.930959 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm22t\" (UniqueName: \"kubernetes.io/projected/a360f605-0014-46ee-973c-48d792d7e895-kube-api-access-dm22t\") on node \"crc\" DevicePath \"\"" Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.930996 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:09:03 crc kubenswrapper[4706]: I1208 20:09:03.931007 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a360f605-0014-46ee-973c-48d792d7e895-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:09:04 crc kubenswrapper[4706]: I1208 20:09:04.647931 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzbqp" Dec 08 20:09:04 crc kubenswrapper[4706]: I1208 20:09:04.703930 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kzbqp"] Dec 08 20:09:04 crc kubenswrapper[4706]: I1208 20:09:04.718751 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kzbqp"] Dec 08 20:09:05 crc kubenswrapper[4706]: I1208 20:09:05.621358 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a360f605-0014-46ee-973c-48d792d7e895" path="/var/lib/kubelet/pods/a360f605-0014-46ee-973c-48d792d7e895/volumes" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.117225 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kpvb7"] Dec 08 20:09:33 crc kubenswrapper[4706]: E1208 20:09:33.118929 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a360f605-0014-46ee-973c-48d792d7e895" containerName="extract-content" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.118951 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a360f605-0014-46ee-973c-48d792d7e895" containerName="extract-content" Dec 08 20:09:33 crc kubenswrapper[4706]: E1208 20:09:33.118995 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a360f605-0014-46ee-973c-48d792d7e895" containerName="registry-server" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.119002 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a360f605-0014-46ee-973c-48d792d7e895" containerName="registry-server" Dec 08 20:09:33 crc kubenswrapper[4706]: E1208 20:09:33.119031 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a360f605-0014-46ee-973c-48d792d7e895" containerName="extract-utilities" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.119043 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a360f605-0014-46ee-973c-48d792d7e895" containerName="extract-utilities" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.119317 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a360f605-0014-46ee-973c-48d792d7e895" containerName="registry-server" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.122053 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.135800 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpvb7"] Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.192626 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8td5\" (UniqueName: \"kubernetes.io/projected/73ae497f-976d-4b85-8693-b76b7c99c397-kube-api-access-j8td5\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.193280 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-utilities\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.193425 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-catalog-content\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.296723 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-utilities\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.297208 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-catalog-content\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.297381 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8td5\" (UniqueName: \"kubernetes.io/projected/73ae497f-976d-4b85-8693-b76b7c99c397-kube-api-access-j8td5\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.297541 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-utilities\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.297825 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-catalog-content\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.324382 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8td5\" (UniqueName: \"kubernetes.io/projected/73ae497f-976d-4b85-8693-b76b7c99c397-kube-api-access-j8td5\") pod \"redhat-marketplace-kpvb7\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:33 crc kubenswrapper[4706]: I1208 20:09:33.480804 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:34 crc kubenswrapper[4706]: I1208 20:09:34.321166 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpvb7"] Dec 08 20:09:35 crc kubenswrapper[4706]: I1208 20:09:35.003101 4706 generic.go:334] "Generic (PLEG): container finished" podID="73ae497f-976d-4b85-8693-b76b7c99c397" containerID="fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520" exitCode=0 Dec 08 20:09:35 crc kubenswrapper[4706]: I1208 20:09:35.003514 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpvb7" event={"ID":"73ae497f-976d-4b85-8693-b76b7c99c397","Type":"ContainerDied","Data":"fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520"} Dec 08 20:09:35 crc kubenswrapper[4706]: I1208 20:09:35.003545 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpvb7" event={"ID":"73ae497f-976d-4b85-8693-b76b7c99c397","Type":"ContainerStarted","Data":"de79b680779b50e69a6532312f138d3fef3d65331e611edab20d36de888de750"} Dec 08 20:09:35 crc kubenswrapper[4706]: I1208 20:09:35.835948 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:09:35 crc kubenswrapper[4706]: I1208 20:09:35.836488 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:09:36 crc kubenswrapper[4706]: I1208 20:09:36.016278 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpvb7" event={"ID":"73ae497f-976d-4b85-8693-b76b7c99c397","Type":"ContainerStarted","Data":"88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe"} Dec 08 20:09:37 crc kubenswrapper[4706]: I1208 20:09:37.030013 4706 generic.go:334] "Generic (PLEG): container finished" podID="73ae497f-976d-4b85-8693-b76b7c99c397" containerID="88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe" exitCode=0 Dec 08 20:09:37 crc kubenswrapper[4706]: I1208 20:09:37.030105 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpvb7" event={"ID":"73ae497f-976d-4b85-8693-b76b7c99c397","Type":"ContainerDied","Data":"88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe"} Dec 08 20:09:38 crc kubenswrapper[4706]: I1208 20:09:38.044721 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpvb7" event={"ID":"73ae497f-976d-4b85-8693-b76b7c99c397","Type":"ContainerStarted","Data":"868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b"} Dec 08 20:09:38 crc kubenswrapper[4706]: I1208 20:09:38.075100 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kpvb7" podStartSLOduration=2.608430941 podStartE2EDuration="5.075073883s" podCreationTimestamp="2025-12-08 20:09:33 +0000 UTC" firstStartedPulling="2025-12-08 20:09:35.005970128 +0000 UTC m=+2857.648171151" lastFinishedPulling="2025-12-08 20:09:37.47261309 +0000 UTC m=+2860.114814093" observedRunningTime="2025-12-08 20:09:38.063762843 +0000 UTC m=+2860.705963846" watchObservedRunningTime="2025-12-08 20:09:38.075073883 +0000 UTC m=+2860.717274886" Dec 08 20:09:43 crc kubenswrapper[4706]: I1208 20:09:43.481678 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:43 crc kubenswrapper[4706]: I1208 20:09:43.482388 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:43 crc kubenswrapper[4706]: I1208 20:09:43.549889 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:44 crc kubenswrapper[4706]: I1208 20:09:44.163049 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:44 crc kubenswrapper[4706]: I1208 20:09:44.223386 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpvb7"] Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.125956 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kpvb7" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" containerName="registry-server" containerID="cri-o://868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b" gracePeriod=2 Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.719626 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.751613 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-utilities\") pod \"73ae497f-976d-4b85-8693-b76b7c99c397\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.751806 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-catalog-content\") pod \"73ae497f-976d-4b85-8693-b76b7c99c397\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.752078 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8td5\" (UniqueName: \"kubernetes.io/projected/73ae497f-976d-4b85-8693-b76b7c99c397-kube-api-access-j8td5\") pod \"73ae497f-976d-4b85-8693-b76b7c99c397\" (UID: \"73ae497f-976d-4b85-8693-b76b7c99c397\") " Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.753571 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-utilities" (OuterVolumeSpecName: "utilities") pod "73ae497f-976d-4b85-8693-b76b7c99c397" (UID: "73ae497f-976d-4b85-8693-b76b7c99c397"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.761609 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73ae497f-976d-4b85-8693-b76b7c99c397-kube-api-access-j8td5" (OuterVolumeSpecName: "kube-api-access-j8td5") pod "73ae497f-976d-4b85-8693-b76b7c99c397" (UID: "73ae497f-976d-4b85-8693-b76b7c99c397"). InnerVolumeSpecName "kube-api-access-j8td5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.776520 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73ae497f-976d-4b85-8693-b76b7c99c397" (UID: "73ae497f-976d-4b85-8693-b76b7c99c397"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.856383 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8td5\" (UniqueName: \"kubernetes.io/projected/73ae497f-976d-4b85-8693-b76b7c99c397-kube-api-access-j8td5\") on node \"crc\" DevicePath \"\"" Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.856438 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:09:46 crc kubenswrapper[4706]: I1208 20:09:46.856450 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ae497f-976d-4b85-8693-b76b7c99c397-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.141168 4706 generic.go:334] "Generic (PLEG): container finished" podID="73ae497f-976d-4b85-8693-b76b7c99c397" containerID="868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b" exitCode=0 Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.141235 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpvb7" event={"ID":"73ae497f-976d-4b85-8693-b76b7c99c397","Type":"ContainerDied","Data":"868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b"} Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.141288 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpvb7" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.142599 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpvb7" event={"ID":"73ae497f-976d-4b85-8693-b76b7c99c397","Type":"ContainerDied","Data":"de79b680779b50e69a6532312f138d3fef3d65331e611edab20d36de888de750"} Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.142622 4706 scope.go:117] "RemoveContainer" containerID="868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.169963 4706 scope.go:117] "RemoveContainer" containerID="88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.187982 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpvb7"] Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.200717 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpvb7"] Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.207092 4706 scope.go:117] "RemoveContainer" containerID="fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.249578 4706 scope.go:117] "RemoveContainer" containerID="868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b" Dec 08 20:09:47 crc kubenswrapper[4706]: E1208 20:09:47.250474 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b\": container with ID starting with 868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b not found: ID does not exist" containerID="868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.250518 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b"} err="failed to get container status \"868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b\": rpc error: code = NotFound desc = could not find container \"868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b\": container with ID starting with 868ee2b9962ade2e6325ceef09edaa89563953bfe1bf9ddc71edfc00e867126b not found: ID does not exist" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.250552 4706 scope.go:117] "RemoveContainer" containerID="88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe" Dec 08 20:09:47 crc kubenswrapper[4706]: E1208 20:09:47.251180 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe\": container with ID starting with 88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe not found: ID does not exist" containerID="88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.251214 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe"} err="failed to get container status \"88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe\": rpc error: code = NotFound desc = could not find container \"88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe\": container with ID starting with 88249bbbd78df9b58bcc75054ed28b163b0d8a3577a3fcaf54a28f0011627dfe not found: ID does not exist" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.251234 4706 scope.go:117] "RemoveContainer" containerID="fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520" Dec 08 20:09:47 crc kubenswrapper[4706]: E1208 20:09:47.251982 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520\": container with ID starting with fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520 not found: ID does not exist" containerID="fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.252013 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520"} err="failed to get container status \"fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520\": rpc error: code = NotFound desc = could not find container \"fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520\": container with ID starting with fe7108e1f9b269bb272c144689649b5010d6dc6cc02fabc53073bbeaadb02520 not found: ID does not exist" Dec 08 20:09:47 crc kubenswrapper[4706]: I1208 20:09:47.620722 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" path="/var/lib/kubelet/pods/73ae497f-976d-4b85-8693-b76b7c99c397/volumes" Dec 08 20:10:05 crc kubenswrapper[4706]: I1208 20:10:05.836400 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:10:05 crc kubenswrapper[4706]: I1208 20:10:05.837196 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:10:35 crc kubenswrapper[4706]: I1208 20:10:35.836533 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:10:35 crc kubenswrapper[4706]: I1208 20:10:35.837172 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:10:35 crc kubenswrapper[4706]: I1208 20:10:35.837281 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 20:10:35 crc kubenswrapper[4706]: I1208 20:10:35.838469 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56b0d08effe4a9c86e0110c858d054b693a2be2056b7c9c944051a267d32813e"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 20:10:35 crc kubenswrapper[4706]: I1208 20:10:35.838557 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://56b0d08effe4a9c86e0110c858d054b693a2be2056b7c9c944051a267d32813e" gracePeriod=600 Dec 08 20:10:36 crc kubenswrapper[4706]: I1208 20:10:36.766186 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="56b0d08effe4a9c86e0110c858d054b693a2be2056b7c9c944051a267d32813e" exitCode=0 Dec 08 20:10:36 crc kubenswrapper[4706]: I1208 20:10:36.766383 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"56b0d08effe4a9c86e0110c858d054b693a2be2056b7c9c944051a267d32813e"} Dec 08 20:10:36 crc kubenswrapper[4706]: I1208 20:10:36.766911 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5"} Dec 08 20:10:36 crc kubenswrapper[4706]: I1208 20:10:36.766934 4706 scope.go:117] "RemoveContainer" containerID="813bb17583caa253569e57ae48784e5cc70fe84a991da498b236c61cb34db77d" Dec 08 20:10:41 crc kubenswrapper[4706]: I1208 20:10:41.839384 4706 generic.go:334] "Generic (PLEG): container finished" podID="39dc4f79-e40f-479c-b82b-8ac18056c9ec" containerID="620febf1a23182cb8e1c72918678b48cd2a244073703e847e6f96fe169b6fd1e" exitCode=0 Dec 08 20:10:41 crc kubenswrapper[4706]: I1208 20:10:41.839473 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" event={"ID":"39dc4f79-e40f-479c-b82b-8ac18056c9ec","Type":"ContainerDied","Data":"620febf1a23182cb8e1c72918678b48cd2a244073703e847e6f96fe169b6fd1e"} Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.449748 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.561163 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-inventory\") pod \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.561408 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-telemetry-combined-ca-bundle\") pod \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.561489 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-2\") pod \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.561627 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ssh-key\") pod \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.561666 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-1\") pod \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.561758 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-0\") pod \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.561844 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nbtt\" (UniqueName: \"kubernetes.io/projected/39dc4f79-e40f-479c-b82b-8ac18056c9ec-kube-api-access-5nbtt\") pod \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\" (UID: \"39dc4f79-e40f-479c-b82b-8ac18056c9ec\") " Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.569187 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "39dc4f79-e40f-479c-b82b-8ac18056c9ec" (UID: "39dc4f79-e40f-479c-b82b-8ac18056c9ec"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.569735 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39dc4f79-e40f-479c-b82b-8ac18056c9ec-kube-api-access-5nbtt" (OuterVolumeSpecName: "kube-api-access-5nbtt") pod "39dc4f79-e40f-479c-b82b-8ac18056c9ec" (UID: "39dc4f79-e40f-479c-b82b-8ac18056c9ec"). InnerVolumeSpecName "kube-api-access-5nbtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.596482 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "39dc4f79-e40f-479c-b82b-8ac18056c9ec" (UID: "39dc4f79-e40f-479c-b82b-8ac18056c9ec"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.597795 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "39dc4f79-e40f-479c-b82b-8ac18056c9ec" (UID: "39dc4f79-e40f-479c-b82b-8ac18056c9ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.599136 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "39dc4f79-e40f-479c-b82b-8ac18056c9ec" (UID: "39dc4f79-e40f-479c-b82b-8ac18056c9ec"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.602667 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-inventory" (OuterVolumeSpecName: "inventory") pod "39dc4f79-e40f-479c-b82b-8ac18056c9ec" (UID: "39dc4f79-e40f-479c-b82b-8ac18056c9ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.620247 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "39dc4f79-e40f-479c-b82b-8ac18056c9ec" (UID: "39dc4f79-e40f-479c-b82b-8ac18056c9ec"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.670161 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.670226 4706 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.670240 4706 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.670467 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nbtt\" (UniqueName: \"kubernetes.io/projected/39dc4f79-e40f-479c-b82b-8ac18056c9ec-kube-api-access-5nbtt\") on node \"crc\" DevicePath \"\"" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.670481 4706 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.670515 4706 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.670527 4706 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/39dc4f79-e40f-479c-b82b-8ac18056c9ec-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.866894 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" event={"ID":"39dc4f79-e40f-479c-b82b-8ac18056c9ec","Type":"ContainerDied","Data":"a32a13e0fbc8bf9d38fe5b9b49c1aabedccb63648a6baefff71d4ff1ed058dfc"} Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.866961 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a32a13e0fbc8bf9d38fe5b9b49c1aabedccb63648a6baefff71d4ff1ed058dfc" Dec 08 20:10:43 crc kubenswrapper[4706]: I1208 20:10:43.867035 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.370871 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dzqr7"] Dec 08 20:11:24 crc kubenswrapper[4706]: E1208 20:11:24.372530 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" containerName="registry-server" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.372549 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" containerName="registry-server" Dec 08 20:11:24 crc kubenswrapper[4706]: E1208 20:11:24.372564 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39dc4f79-e40f-479c-b82b-8ac18056c9ec" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.372576 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="39dc4f79-e40f-479c-b82b-8ac18056c9ec" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 08 20:11:24 crc kubenswrapper[4706]: E1208 20:11:24.372598 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" containerName="extract-utilities" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.372607 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" containerName="extract-utilities" Dec 08 20:11:24 crc kubenswrapper[4706]: E1208 20:11:24.372632 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" containerName="extract-content" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.372641 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" containerName="extract-content" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.372959 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="39dc4f79-e40f-479c-b82b-8ac18056c9ec" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.372995 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="73ae497f-976d-4b85-8693-b76b7c99c397" containerName="registry-server" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.375578 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.383488 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dzqr7"] Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.573111 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtqmk\" (UniqueName: \"kubernetes.io/projected/4995abcf-95fd-4231-a5ef-c081a05b225d-kube-api-access-wtqmk\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.573872 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-utilities\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.574388 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-catalog-content\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.676850 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-catalog-content\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.677802 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtqmk\" (UniqueName: \"kubernetes.io/projected/4995abcf-95fd-4231-a5ef-c081a05b225d-kube-api-access-wtqmk\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.678091 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-catalog-content\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.678159 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-utilities\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.678761 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-utilities\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:24 crc kubenswrapper[4706]: I1208 20:11:24.717514 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtqmk\" (UniqueName: \"kubernetes.io/projected/4995abcf-95fd-4231-a5ef-c081a05b225d-kube-api-access-wtqmk\") pod \"community-operators-dzqr7\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:25 crc kubenswrapper[4706]: I1208 20:11:25.013853 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:25 crc kubenswrapper[4706]: I1208 20:11:25.585557 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dzqr7"] Dec 08 20:11:26 crc kubenswrapper[4706]: I1208 20:11:26.329791 4706 generic.go:334] "Generic (PLEG): container finished" podID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerID="9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271" exitCode=0 Dec 08 20:11:26 crc kubenswrapper[4706]: I1208 20:11:26.329901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzqr7" event={"ID":"4995abcf-95fd-4231-a5ef-c081a05b225d","Type":"ContainerDied","Data":"9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271"} Dec 08 20:11:26 crc kubenswrapper[4706]: I1208 20:11:26.330182 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzqr7" event={"ID":"4995abcf-95fd-4231-a5ef-c081a05b225d","Type":"ContainerStarted","Data":"3d27fa554c6cefcdcfaa846487cc4ef9ca69df46a7899857e694cd43ed7cf822"} Dec 08 20:11:26 crc kubenswrapper[4706]: I1208 20:11:26.332671 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 20:11:27 crc kubenswrapper[4706]: I1208 20:11:27.345251 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzqr7" event={"ID":"4995abcf-95fd-4231-a5ef-c081a05b225d","Type":"ContainerStarted","Data":"9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6"} Dec 08 20:11:28 crc kubenswrapper[4706]: I1208 20:11:28.359901 4706 generic.go:334] "Generic (PLEG): container finished" podID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerID="9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6" exitCode=0 Dec 08 20:11:28 crc kubenswrapper[4706]: I1208 20:11:28.359959 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzqr7" event={"ID":"4995abcf-95fd-4231-a5ef-c081a05b225d","Type":"ContainerDied","Data":"9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6"} Dec 08 20:11:29 crc kubenswrapper[4706]: I1208 20:11:29.372427 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzqr7" event={"ID":"4995abcf-95fd-4231-a5ef-c081a05b225d","Type":"ContainerStarted","Data":"bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190"} Dec 08 20:11:29 crc kubenswrapper[4706]: I1208 20:11:29.410428 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dzqr7" podStartSLOduration=2.993421994 podStartE2EDuration="5.410407568s" podCreationTimestamp="2025-12-08 20:11:24 +0000 UTC" firstStartedPulling="2025-12-08 20:11:26.332357779 +0000 UTC m=+2968.974558792" lastFinishedPulling="2025-12-08 20:11:28.749343363 +0000 UTC m=+2971.391544366" observedRunningTime="2025-12-08 20:11:29.407105514 +0000 UTC m=+2972.049306517" watchObservedRunningTime="2025-12-08 20:11:29.410407568 +0000 UTC m=+2972.052608571" Dec 08 20:11:35 crc kubenswrapper[4706]: I1208 20:11:35.013994 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:35 crc kubenswrapper[4706]: I1208 20:11:35.014828 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:35 crc kubenswrapper[4706]: I1208 20:11:35.073651 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:35 crc kubenswrapper[4706]: I1208 20:11:35.501411 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:35 crc kubenswrapper[4706]: I1208 20:11:35.561137 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dzqr7"] Dec 08 20:11:37 crc kubenswrapper[4706]: I1208 20:11:37.466538 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dzqr7" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerName="registry-server" containerID="cri-o://bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190" gracePeriod=2 Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.182691 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.241668 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtqmk\" (UniqueName: \"kubernetes.io/projected/4995abcf-95fd-4231-a5ef-c081a05b225d-kube-api-access-wtqmk\") pod \"4995abcf-95fd-4231-a5ef-c081a05b225d\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.242066 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-catalog-content\") pod \"4995abcf-95fd-4231-a5ef-c081a05b225d\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.242133 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-utilities\") pod \"4995abcf-95fd-4231-a5ef-c081a05b225d\" (UID: \"4995abcf-95fd-4231-a5ef-c081a05b225d\") " Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.243355 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-utilities" (OuterVolumeSpecName: "utilities") pod "4995abcf-95fd-4231-a5ef-c081a05b225d" (UID: "4995abcf-95fd-4231-a5ef-c081a05b225d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.255591 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4995abcf-95fd-4231-a5ef-c081a05b225d-kube-api-access-wtqmk" (OuterVolumeSpecName: "kube-api-access-wtqmk") pod "4995abcf-95fd-4231-a5ef-c081a05b225d" (UID: "4995abcf-95fd-4231-a5ef-c081a05b225d"). InnerVolumeSpecName "kube-api-access-wtqmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.303323 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4995abcf-95fd-4231-a5ef-c081a05b225d" (UID: "4995abcf-95fd-4231-a5ef-c081a05b225d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.345322 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.345695 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4995abcf-95fd-4231-a5ef-c081a05b225d-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.345788 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtqmk\" (UniqueName: \"kubernetes.io/projected/4995abcf-95fd-4231-a5ef-c081a05b225d-kube-api-access-wtqmk\") on node \"crc\" DevicePath \"\"" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.479364 4706 generic.go:334] "Generic (PLEG): container finished" podID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerID="bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190" exitCode=0 Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.479419 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzqr7" event={"ID":"4995abcf-95fd-4231-a5ef-c081a05b225d","Type":"ContainerDied","Data":"bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190"} Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.479458 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzqr7" event={"ID":"4995abcf-95fd-4231-a5ef-c081a05b225d","Type":"ContainerDied","Data":"3d27fa554c6cefcdcfaa846487cc4ef9ca69df46a7899857e694cd43ed7cf822"} Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.479480 4706 scope.go:117] "RemoveContainer" containerID="bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.479489 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzqr7" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.508162 4706 scope.go:117] "RemoveContainer" containerID="9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.520777 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dzqr7"] Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.532759 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dzqr7"] Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.548601 4706 scope.go:117] "RemoveContainer" containerID="9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.586115 4706 scope.go:117] "RemoveContainer" containerID="bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190" Dec 08 20:11:38 crc kubenswrapper[4706]: E1208 20:11:38.587598 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190\": container with ID starting with bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190 not found: ID does not exist" containerID="bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.587654 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190"} err="failed to get container status \"bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190\": rpc error: code = NotFound desc = could not find container \"bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190\": container with ID starting with bc40ac3d4091df3139caa23cbd0fedc55292a53e21fa6abaffe529fdc2fe5190 not found: ID does not exist" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.587696 4706 scope.go:117] "RemoveContainer" containerID="9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6" Dec 08 20:11:38 crc kubenswrapper[4706]: E1208 20:11:38.588112 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6\": container with ID starting with 9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6 not found: ID does not exist" containerID="9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.588144 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6"} err="failed to get container status \"9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6\": rpc error: code = NotFound desc = could not find container \"9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6\": container with ID starting with 9eafe1e67b23fa7eb70c8848bc1051b068d4f7bb7bbe7e3e26b2e2d2eaabd0a6 not found: ID does not exist" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.588162 4706 scope.go:117] "RemoveContainer" containerID="9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271" Dec 08 20:11:38 crc kubenswrapper[4706]: E1208 20:11:38.588624 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271\": container with ID starting with 9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271 not found: ID does not exist" containerID="9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271" Dec 08 20:11:38 crc kubenswrapper[4706]: I1208 20:11:38.588684 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271"} err="failed to get container status \"9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271\": rpc error: code = NotFound desc = could not find container \"9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271\": container with ID starting with 9bdbe1dd10f67b5a6b7c7a45b5836661b8444ed955f6fb87f58c4836344ff271 not found: ID does not exist" Dec 08 20:11:39 crc kubenswrapper[4706]: I1208 20:11:39.623056 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" path="/var/lib/kubelet/pods/4995abcf-95fd-4231-a5ef-c081a05b225d/volumes" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.861212 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 08 20:12:03 crc kubenswrapper[4706]: E1208 20:12:03.862385 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerName="registry-server" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.862401 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerName="registry-server" Dec 08 20:12:03 crc kubenswrapper[4706]: E1208 20:12:03.862449 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerName="extract-utilities" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.862458 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerName="extract-utilities" Dec 08 20:12:03 crc kubenswrapper[4706]: E1208 20:12:03.862467 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerName="extract-content" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.862474 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerName="extract-content" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.862693 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4995abcf-95fd-4231-a5ef-c081a05b225d" containerName="registry-server" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.863577 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.866936 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.867131 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tg4ng" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.870056 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.870399 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.885771 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.965884 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.965978 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.966221 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-config-data\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.966351 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.966433 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.966880 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.966956 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.967020 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:03 crc kubenswrapper[4706]: I1208 20:12:03.967324 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmrcs\" (UniqueName: \"kubernetes.io/projected/e673aa84-089b-4ab2-abcd-9c3827d97cbf-kube-api-access-hmrcs\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.070288 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.070388 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.070455 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-config-data\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.070481 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.070524 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.070661 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.071127 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.071185 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.070754 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.071315 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.071553 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.071842 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.072953 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-config-data\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.073211 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmrcs\" (UniqueName: \"kubernetes.io/projected/e673aa84-089b-4ab2-abcd-9c3827d97cbf-kube-api-access-hmrcs\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.079539 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.079552 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.080970 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.096968 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmrcs\" (UniqueName: \"kubernetes.io/projected/e673aa84-089b-4ab2-abcd-9c3827d97cbf-kube-api-access-hmrcs\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.120640 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.187988 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.702308 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 08 20:12:04 crc kubenswrapper[4706]: I1208 20:12:04.780734 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"e673aa84-089b-4ab2-abcd-9c3827d97cbf","Type":"ContainerStarted","Data":"216d0263ba3354d4bc5a9b0f68f4768e1274639495deda9908d4b5ee192b1e7b"} Dec 08 20:12:41 crc kubenswrapper[4706]: E1208 20:12:41.831770 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 08 20:12:41 crc kubenswrapper[4706]: E1208 20:12:41.832749 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hmrcs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(e673aa84-089b-4ab2-abcd-9c3827d97cbf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 20:12:41 crc kubenswrapper[4706]: E1208 20:12:41.834011 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="e673aa84-089b-4ab2-abcd-9c3827d97cbf" Dec 08 20:12:42 crc kubenswrapper[4706]: E1208 20:12:42.285120 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="e673aa84-089b-4ab2-abcd-9c3827d97cbf" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.099037 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vrf2v"] Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.103675 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.111869 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vrf2v"] Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.274592 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzw4g\" (UniqueName: \"kubernetes.io/projected/a77cc335-0144-440e-ab07-2b314e4102b9-kube-api-access-gzw4g\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.274691 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-utilities\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.274773 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-catalog-content\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.378058 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-utilities\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.378580 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-catalog-content\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.378767 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzw4g\" (UniqueName: \"kubernetes.io/projected/a77cc335-0144-440e-ab07-2b314e4102b9-kube-api-access-gzw4g\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.379354 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-catalog-content\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.379531 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-utilities\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.403317 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzw4g\" (UniqueName: \"kubernetes.io/projected/a77cc335-0144-440e-ab07-2b314e4102b9-kube-api-access-gzw4g\") pod \"redhat-operators-vrf2v\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.449695 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:43 crc kubenswrapper[4706]: I1208 20:12:43.985896 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vrf2v"] Dec 08 20:12:43 crc kubenswrapper[4706]: W1208 20:12:43.991486 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda77cc335_0144_440e_ab07_2b314e4102b9.slice/crio-8dcbd2fa8027aa4529b8fe84b056bbd42bb3915f72a2ebb8acbc9b19a3cac8c1 WatchSource:0}: Error finding container 8dcbd2fa8027aa4529b8fe84b056bbd42bb3915f72a2ebb8acbc9b19a3cac8c1: Status 404 returned error can't find the container with id 8dcbd2fa8027aa4529b8fe84b056bbd42bb3915f72a2ebb8acbc9b19a3cac8c1 Dec 08 20:12:44 crc kubenswrapper[4706]: I1208 20:12:44.320310 4706 generic.go:334] "Generic (PLEG): container finished" podID="a77cc335-0144-440e-ab07-2b314e4102b9" containerID="489bf9247ee280ec31fb597909266f4073da4739363662e8985e76c144742f62" exitCode=0 Dec 08 20:12:44 crc kubenswrapper[4706]: I1208 20:12:44.320376 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrf2v" event={"ID":"a77cc335-0144-440e-ab07-2b314e4102b9","Type":"ContainerDied","Data":"489bf9247ee280ec31fb597909266f4073da4739363662e8985e76c144742f62"} Dec 08 20:12:44 crc kubenswrapper[4706]: I1208 20:12:44.320424 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrf2v" event={"ID":"a77cc335-0144-440e-ab07-2b314e4102b9","Type":"ContainerStarted","Data":"8dcbd2fa8027aa4529b8fe84b056bbd42bb3915f72a2ebb8acbc9b19a3cac8c1"} Dec 08 20:12:45 crc kubenswrapper[4706]: I1208 20:12:45.333968 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrf2v" event={"ID":"a77cc335-0144-440e-ab07-2b314e4102b9","Type":"ContainerStarted","Data":"db8b2c62b4735292a425e6a4f724668a5be2a6db235bbf3756a2306d68dee85e"} Dec 08 20:12:49 crc kubenswrapper[4706]: I1208 20:12:49.381961 4706 generic.go:334] "Generic (PLEG): container finished" podID="a77cc335-0144-440e-ab07-2b314e4102b9" containerID="db8b2c62b4735292a425e6a4f724668a5be2a6db235bbf3756a2306d68dee85e" exitCode=0 Dec 08 20:12:49 crc kubenswrapper[4706]: I1208 20:12:49.382055 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrf2v" event={"ID":"a77cc335-0144-440e-ab07-2b314e4102b9","Type":"ContainerDied","Data":"db8b2c62b4735292a425e6a4f724668a5be2a6db235bbf3756a2306d68dee85e"} Dec 08 20:12:50 crc kubenswrapper[4706]: I1208 20:12:50.399109 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrf2v" event={"ID":"a77cc335-0144-440e-ab07-2b314e4102b9","Type":"ContainerStarted","Data":"4fdcd17c9b33304272e48b5f10a9e9a730bdf0cf0e85cdd19392f7ecce4d18a9"} Dec 08 20:12:50 crc kubenswrapper[4706]: I1208 20:12:50.426375 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vrf2v" podStartSLOduration=1.8124774399999999 podStartE2EDuration="7.426350091s" podCreationTimestamp="2025-12-08 20:12:43 +0000 UTC" firstStartedPulling="2025-12-08 20:12:44.32299561 +0000 UTC m=+3046.965196613" lastFinishedPulling="2025-12-08 20:12:49.936868261 +0000 UTC m=+3052.579069264" observedRunningTime="2025-12-08 20:12:50.425021144 +0000 UTC m=+3053.067222147" watchObservedRunningTime="2025-12-08 20:12:50.426350091 +0000 UTC m=+3053.068551094" Dec 08 20:12:53 crc kubenswrapper[4706]: I1208 20:12:53.449926 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:53 crc kubenswrapper[4706]: I1208 20:12:53.450690 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:12:54 crc kubenswrapper[4706]: I1208 20:12:54.503211 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vrf2v" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="registry-server" probeResult="failure" output=< Dec 08 20:12:54 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 20:12:54 crc kubenswrapper[4706]: > Dec 08 20:12:58 crc kubenswrapper[4706]: I1208 20:12:58.492794 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"e673aa84-089b-4ab2-abcd-9c3827d97cbf","Type":"ContainerStarted","Data":"f9fe5d649830e7d88cfd8ce01f0476b5fbc093305e4c379ff7505a1abb1f291d"} Dec 08 20:12:58 crc kubenswrapper[4706]: I1208 20:12:58.524073 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.222613886 podStartE2EDuration="56.524048313s" podCreationTimestamp="2025-12-08 20:12:02 +0000 UTC" firstStartedPulling="2025-12-08 20:12:04.717401264 +0000 UTC m=+3007.359602267" lastFinishedPulling="2025-12-08 20:12:57.018835691 +0000 UTC m=+3059.661036694" observedRunningTime="2025-12-08 20:12:58.515372442 +0000 UTC m=+3061.157573485" watchObservedRunningTime="2025-12-08 20:12:58.524048313 +0000 UTC m=+3061.166249316" Dec 08 20:13:03 crc kubenswrapper[4706]: I1208 20:13:03.517651 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:13:03 crc kubenswrapper[4706]: I1208 20:13:03.578867 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:13:03 crc kubenswrapper[4706]: I1208 20:13:03.766369 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vrf2v"] Dec 08 20:13:04 crc kubenswrapper[4706]: I1208 20:13:04.556694 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vrf2v" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="registry-server" containerID="cri-o://4fdcd17c9b33304272e48b5f10a9e9a730bdf0cf0e85cdd19392f7ecce4d18a9" gracePeriod=2 Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.573373 4706 generic.go:334] "Generic (PLEG): container finished" podID="a77cc335-0144-440e-ab07-2b314e4102b9" containerID="4fdcd17c9b33304272e48b5f10a9e9a730bdf0cf0e85cdd19392f7ecce4d18a9" exitCode=0 Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.573441 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrf2v" event={"ID":"a77cc335-0144-440e-ab07-2b314e4102b9","Type":"ContainerDied","Data":"4fdcd17c9b33304272e48b5f10a9e9a730bdf0cf0e85cdd19392f7ecce4d18a9"} Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.573795 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrf2v" event={"ID":"a77cc335-0144-440e-ab07-2b314e4102b9","Type":"ContainerDied","Data":"8dcbd2fa8027aa4529b8fe84b056bbd42bb3915f72a2ebb8acbc9b19a3cac8c1"} Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.573809 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dcbd2fa8027aa4529b8fe84b056bbd42bb3915f72a2ebb8acbc9b19a3cac8c1" Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.659081 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.761123 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzw4g\" (UniqueName: \"kubernetes.io/projected/a77cc335-0144-440e-ab07-2b314e4102b9-kube-api-access-gzw4g\") pod \"a77cc335-0144-440e-ab07-2b314e4102b9\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.761440 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-utilities\") pod \"a77cc335-0144-440e-ab07-2b314e4102b9\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.761476 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-catalog-content\") pod \"a77cc335-0144-440e-ab07-2b314e4102b9\" (UID: \"a77cc335-0144-440e-ab07-2b314e4102b9\") " Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.762738 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-utilities" (OuterVolumeSpecName: "utilities") pod "a77cc335-0144-440e-ab07-2b314e4102b9" (UID: "a77cc335-0144-440e-ab07-2b314e4102b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.770291 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a77cc335-0144-440e-ab07-2b314e4102b9-kube-api-access-gzw4g" (OuterVolumeSpecName: "kube-api-access-gzw4g") pod "a77cc335-0144-440e-ab07-2b314e4102b9" (UID: "a77cc335-0144-440e-ab07-2b314e4102b9"). InnerVolumeSpecName "kube-api-access-gzw4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.836551 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.837091 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.865380 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzw4g\" (UniqueName: \"kubernetes.io/projected/a77cc335-0144-440e-ab07-2b314e4102b9-kube-api-access-gzw4g\") on node \"crc\" DevicePath \"\"" Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.865427 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.937079 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a77cc335-0144-440e-ab07-2b314e4102b9" (UID: "a77cc335-0144-440e-ab07-2b314e4102b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:13:05 crc kubenswrapper[4706]: I1208 20:13:05.967540 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a77cc335-0144-440e-ab07-2b314e4102b9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:13:06 crc kubenswrapper[4706]: I1208 20:13:06.582998 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrf2v" Dec 08 20:13:06 crc kubenswrapper[4706]: I1208 20:13:06.634156 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vrf2v"] Dec 08 20:13:06 crc kubenswrapper[4706]: I1208 20:13:06.647175 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vrf2v"] Dec 08 20:13:07 crc kubenswrapper[4706]: I1208 20:13:07.636388 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" path="/var/lib/kubelet/pods/a77cc335-0144-440e-ab07-2b314e4102b9/volumes" Dec 08 20:13:35 crc kubenswrapper[4706]: I1208 20:13:35.836443 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:13:35 crc kubenswrapper[4706]: I1208 20:13:35.837235 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:14:05 crc kubenswrapper[4706]: I1208 20:14:05.836324 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:14:05 crc kubenswrapper[4706]: I1208 20:14:05.837280 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:14:05 crc kubenswrapper[4706]: I1208 20:14:05.837367 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 20:14:05 crc kubenswrapper[4706]: I1208 20:14:05.838549 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 20:14:05 crc kubenswrapper[4706]: I1208 20:14:05.838618 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" gracePeriod=600 Dec 08 20:14:05 crc kubenswrapper[4706]: E1208 20:14:05.964305 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:14:06 crc kubenswrapper[4706]: I1208 20:14:06.342558 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" exitCode=0 Dec 08 20:14:06 crc kubenswrapper[4706]: I1208 20:14:06.342645 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5"} Dec 08 20:14:06 crc kubenswrapper[4706]: I1208 20:14:06.342746 4706 scope.go:117] "RemoveContainer" containerID="56b0d08effe4a9c86e0110c858d054b693a2be2056b7c9c944051a267d32813e" Dec 08 20:14:06 crc kubenswrapper[4706]: I1208 20:14:06.343840 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:14:06 crc kubenswrapper[4706]: E1208 20:14:06.344250 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:14:17 crc kubenswrapper[4706]: I1208 20:14:17.648104 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:14:17 crc kubenswrapper[4706]: E1208 20:14:17.650087 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:14:28 crc kubenswrapper[4706]: I1208 20:14:28.610427 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:14:28 crc kubenswrapper[4706]: E1208 20:14:28.611942 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:14:39 crc kubenswrapper[4706]: I1208 20:14:39.608814 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:14:39 crc kubenswrapper[4706]: E1208 20:14:39.609923 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:14:50 crc kubenswrapper[4706]: I1208 20:14:50.609092 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:14:50 crc kubenswrapper[4706]: E1208 20:14:50.609969 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.182228 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl"] Dec 08 20:15:00 crc kubenswrapper[4706]: E1208 20:15:00.185005 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="extract-content" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.185096 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="extract-content" Dec 08 20:15:00 crc kubenswrapper[4706]: E1208 20:15:00.185157 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="extract-utilities" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.185181 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="extract-utilities" Dec 08 20:15:00 crc kubenswrapper[4706]: E1208 20:15:00.185242 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="registry-server" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.185251 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="registry-server" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.185806 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a77cc335-0144-440e-ab07-2b314e4102b9" containerName="registry-server" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.187574 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.192431 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.192463 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.196415 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl"] Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.320878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0df0a0fe-f626-4a63-b187-678e51d1b979-config-volume\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.320962 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f72q\" (UniqueName: \"kubernetes.io/projected/0df0a0fe-f626-4a63-b187-678e51d1b979-kube-api-access-4f72q\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.321230 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0df0a0fe-f626-4a63-b187-678e51d1b979-secret-volume\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.424769 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0df0a0fe-f626-4a63-b187-678e51d1b979-config-volume\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.424887 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f72q\" (UniqueName: \"kubernetes.io/projected/0df0a0fe-f626-4a63-b187-678e51d1b979-kube-api-access-4f72q\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.425202 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0df0a0fe-f626-4a63-b187-678e51d1b979-secret-volume\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.428023 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0df0a0fe-f626-4a63-b187-678e51d1b979-config-volume\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.447921 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0df0a0fe-f626-4a63-b187-678e51d1b979-secret-volume\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.460181 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f72q\" (UniqueName: \"kubernetes.io/projected/0df0a0fe-f626-4a63-b187-678e51d1b979-kube-api-access-4f72q\") pod \"collect-profiles-29420415-r6ffl\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:00 crc kubenswrapper[4706]: I1208 20:15:00.520705 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:01 crc kubenswrapper[4706]: I1208 20:15:01.096250 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl"] Dec 08 20:15:02 crc kubenswrapper[4706]: I1208 20:15:02.043921 4706 generic.go:334] "Generic (PLEG): container finished" podID="0df0a0fe-f626-4a63-b187-678e51d1b979" containerID="c25f0e416d95747d1bb15b92392289527ca58325e3c55dd62294622f3ae2dde4" exitCode=0 Dec 08 20:15:02 crc kubenswrapper[4706]: I1208 20:15:02.044048 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" event={"ID":"0df0a0fe-f626-4a63-b187-678e51d1b979","Type":"ContainerDied","Data":"c25f0e416d95747d1bb15b92392289527ca58325e3c55dd62294622f3ae2dde4"} Dec 08 20:15:02 crc kubenswrapper[4706]: I1208 20:15:02.044372 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" event={"ID":"0df0a0fe-f626-4a63-b187-678e51d1b979","Type":"ContainerStarted","Data":"079b2fe0e254e75bc15785f5af5b4dfcdd39e141bb329dc3f7be5bf783bb361e"} Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.608793 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:15:03 crc kubenswrapper[4706]: E1208 20:15:03.609544 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.711303 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.817004 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0df0a0fe-f626-4a63-b187-678e51d1b979-config-volume\") pod \"0df0a0fe-f626-4a63-b187-678e51d1b979\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.817119 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0df0a0fe-f626-4a63-b187-678e51d1b979-secret-volume\") pod \"0df0a0fe-f626-4a63-b187-678e51d1b979\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.818023 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0df0a0fe-f626-4a63-b187-678e51d1b979-config-volume" (OuterVolumeSpecName: "config-volume") pod "0df0a0fe-f626-4a63-b187-678e51d1b979" (UID: "0df0a0fe-f626-4a63-b187-678e51d1b979"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.818700 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f72q\" (UniqueName: \"kubernetes.io/projected/0df0a0fe-f626-4a63-b187-678e51d1b979-kube-api-access-4f72q\") pod \"0df0a0fe-f626-4a63-b187-678e51d1b979\" (UID: \"0df0a0fe-f626-4a63-b187-678e51d1b979\") " Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.819373 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0df0a0fe-f626-4a63-b187-678e51d1b979-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.829241 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0df0a0fe-f626-4a63-b187-678e51d1b979-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0df0a0fe-f626-4a63-b187-678e51d1b979" (UID: "0df0a0fe-f626-4a63-b187-678e51d1b979"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.830109 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0df0a0fe-f626-4a63-b187-678e51d1b979-kube-api-access-4f72q" (OuterVolumeSpecName: "kube-api-access-4f72q") pod "0df0a0fe-f626-4a63-b187-678e51d1b979" (UID: "0df0a0fe-f626-4a63-b187-678e51d1b979"). InnerVolumeSpecName "kube-api-access-4f72q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.921784 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f72q\" (UniqueName: \"kubernetes.io/projected/0df0a0fe-f626-4a63-b187-678e51d1b979-kube-api-access-4f72q\") on node \"crc\" DevicePath \"\"" Dec 08 20:15:03 crc kubenswrapper[4706]: I1208 20:15:03.921837 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0df0a0fe-f626-4a63-b187-678e51d1b979-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 20:15:04 crc kubenswrapper[4706]: I1208 20:15:04.071196 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" event={"ID":"0df0a0fe-f626-4a63-b187-678e51d1b979","Type":"ContainerDied","Data":"079b2fe0e254e75bc15785f5af5b4dfcdd39e141bb329dc3f7be5bf783bb361e"} Dec 08 20:15:04 crc kubenswrapper[4706]: I1208 20:15:04.071292 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="079b2fe0e254e75bc15785f5af5b4dfcdd39e141bb329dc3f7be5bf783bb361e" Dec 08 20:15:04 crc kubenswrapper[4706]: I1208 20:15:04.071385 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420415-r6ffl" Dec 08 20:15:04 crc kubenswrapper[4706]: I1208 20:15:04.831216 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b"] Dec 08 20:15:04 crc kubenswrapper[4706]: I1208 20:15:04.846362 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420370-clr2b"] Dec 08 20:15:05 crc kubenswrapper[4706]: I1208 20:15:05.628230 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c86e8022-6d78-4826-aa6e-55fd4c38bf4f" path="/var/lib/kubelet/pods/c86e8022-6d78-4826-aa6e-55fd4c38bf4f/volumes" Dec 08 20:15:13 crc kubenswrapper[4706]: I1208 20:15:13.095694 4706 scope.go:117] "RemoveContainer" containerID="6b32975cff580efe0937b8686c83949688035d63341c5fbfb7bb778ecd5debe0" Dec 08 20:15:13 crc kubenswrapper[4706]: I1208 20:15:13.150988 4706 scope.go:117] "RemoveContainer" containerID="28c2fe18dcf2704258987dd130093c4dd66e3b297b1deca1052b996bda5f2cbd" Dec 08 20:15:13 crc kubenswrapper[4706]: I1208 20:15:13.233034 4706 scope.go:117] "RemoveContainer" containerID="3090c530d69bc6e68720904dd8e0344b23546076e43458ad6254aaf6c8c48b52" Dec 08 20:15:13 crc kubenswrapper[4706]: I1208 20:15:13.277772 4706 scope.go:117] "RemoveContainer" containerID="7f30bbcd9be30b2edfdb6933f5f463e5023cd57a2c77ad078dab76914e216176" Dec 08 20:15:18 crc kubenswrapper[4706]: I1208 20:15:18.609588 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:15:18 crc kubenswrapper[4706]: E1208 20:15:18.610431 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:15:30 crc kubenswrapper[4706]: I1208 20:15:30.608693 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:15:30 crc kubenswrapper[4706]: E1208 20:15:30.609937 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:15:43 crc kubenswrapper[4706]: I1208 20:15:43.610735 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:15:43 crc kubenswrapper[4706]: E1208 20:15:43.612157 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:15:56 crc kubenswrapper[4706]: I1208 20:15:56.608720 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:15:56 crc kubenswrapper[4706]: E1208 20:15:56.610140 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:16:07 crc kubenswrapper[4706]: I1208 20:16:07.608924 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:16:07 crc kubenswrapper[4706]: E1208 20:16:07.610042 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:16:18 crc kubenswrapper[4706]: I1208 20:16:18.608655 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:16:18 crc kubenswrapper[4706]: E1208 20:16:18.609906 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:16:32 crc kubenswrapper[4706]: I1208 20:16:32.608367 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:16:32 crc kubenswrapper[4706]: E1208 20:16:32.609621 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:16:47 crc kubenswrapper[4706]: I1208 20:16:47.620166 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:16:47 crc kubenswrapper[4706]: E1208 20:16:47.621254 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:17:01 crc kubenswrapper[4706]: I1208 20:17:01.609289 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:17:01 crc kubenswrapper[4706]: E1208 20:17:01.610517 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:17:15 crc kubenswrapper[4706]: I1208 20:17:15.608526 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:17:15 crc kubenswrapper[4706]: E1208 20:17:15.609456 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:17:26 crc kubenswrapper[4706]: I1208 20:17:26.608557 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:17:26 crc kubenswrapper[4706]: E1208 20:17:26.609874 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:17:39 crc kubenswrapper[4706]: I1208 20:17:39.609955 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:17:39 crc kubenswrapper[4706]: E1208 20:17:39.612698 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:17:52 crc kubenswrapper[4706]: I1208 20:17:52.609741 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:17:52 crc kubenswrapper[4706]: E1208 20:17:52.611127 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:18:06 crc kubenswrapper[4706]: I1208 20:18:06.609603 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:18:06 crc kubenswrapper[4706]: E1208 20:18:06.610886 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:18:20 crc kubenswrapper[4706]: I1208 20:18:20.608372 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:18:20 crc kubenswrapper[4706]: E1208 20:18:20.609549 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:18:31 crc kubenswrapper[4706]: I1208 20:18:31.609134 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:18:31 crc kubenswrapper[4706]: E1208 20:18:31.610636 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:18:42 crc kubenswrapper[4706]: I1208 20:18:42.609481 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:18:42 crc kubenswrapper[4706]: E1208 20:18:42.610743 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:18:54 crc kubenswrapper[4706]: I1208 20:18:54.609236 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:18:54 crc kubenswrapper[4706]: E1208 20:18:54.610616 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:19:08 crc kubenswrapper[4706]: I1208 20:19:08.610909 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:19:09 crc kubenswrapper[4706]: I1208 20:19:09.133576 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"ae3b6b90e663cc0e30fdffcaf7a2453642e4c1963568b4eafa4406e240816b79"} Dec 08 20:19:13 crc kubenswrapper[4706]: I1208 20:19:13.474233 4706 scope.go:117] "RemoveContainer" containerID="db8b2c62b4735292a425e6a4f724668a5be2a6db235bbf3756a2306d68dee85e" Dec 08 20:19:13 crc kubenswrapper[4706]: I1208 20:19:13.524616 4706 scope.go:117] "RemoveContainer" containerID="489bf9247ee280ec31fb597909266f4073da4739363662e8985e76c144742f62" Dec 08 20:19:13 crc kubenswrapper[4706]: I1208 20:19:13.597067 4706 scope.go:117] "RemoveContainer" containerID="4fdcd17c9b33304272e48b5f10a9e9a730bdf0cf0e85cdd19392f7ecce4d18a9" Dec 08 20:19:30 crc kubenswrapper[4706]: I1208 20:19:30.382735 4706 generic.go:334] "Generic (PLEG): container finished" podID="e673aa84-089b-4ab2-abcd-9c3827d97cbf" containerID="f9fe5d649830e7d88cfd8ce01f0476b5fbc093305e4c379ff7505a1abb1f291d" exitCode=0 Dec 08 20:19:30 crc kubenswrapper[4706]: I1208 20:19:30.383398 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"e673aa84-089b-4ab2-abcd-9c3827d97cbf","Type":"ContainerDied","Data":"f9fe5d649830e7d88cfd8ce01f0476b5fbc093305e4c379ff7505a1abb1f291d"} Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.110801 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273195 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ca-certs\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273319 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ssh-key\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273411 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config-secret\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273490 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-workdir\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273556 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273581 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-config-data\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273683 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmrcs\" (UniqueName: \"kubernetes.io/projected/e673aa84-089b-4ab2-abcd-9c3827d97cbf-kube-api-access-hmrcs\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273801 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.273829 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-temporary\") pod \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\" (UID: \"e673aa84-089b-4ab2-abcd-9c3827d97cbf\") " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.275424 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.282124 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-config-data" (OuterVolumeSpecName: "config-data") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.282436 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e673aa84-089b-4ab2-abcd-9c3827d97cbf-kube-api-access-hmrcs" (OuterVolumeSpecName: "kube-api-access-hmrcs") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "kube-api-access-hmrcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.285018 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.317944 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.325779 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.344427 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.371134 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.377172 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.377217 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e673aa84-089b-4ab2-abcd-9c3827d97cbf-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.377229 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmrcs\" (UniqueName: \"kubernetes.io/projected/e673aa84-089b-4ab2-abcd-9c3827d97cbf-kube-api-access-hmrcs\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.377305 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.377317 4706 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.377330 4706 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.377340 4706 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.377351 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e673aa84-089b-4ab2-abcd-9c3827d97cbf-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.410111 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.412018 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"e673aa84-089b-4ab2-abcd-9c3827d97cbf","Type":"ContainerDied","Data":"216d0263ba3354d4bc5a9b0f68f4768e1274639495deda9908d4b5ee192b1e7b"} Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.412065 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216d0263ba3354d4bc5a9b0f68f4768e1274639495deda9908d4b5ee192b1e7b" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.412139 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.480657 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.734032 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "e673aa84-089b-4ab2-abcd-9c3827d97cbf" (UID: "e673aa84-089b-4ab2-abcd-9c3827d97cbf"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:19:32 crc kubenswrapper[4706]: I1208 20:19:32.788966 4706 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e673aa84-089b-4ab2-abcd-9c3827d97cbf-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.622206 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 08 20:19:37 crc kubenswrapper[4706]: E1208 20:19:37.623896 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e673aa84-089b-4ab2-abcd-9c3827d97cbf" containerName="tempest-tests-tempest-tests-runner" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.623921 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e673aa84-089b-4ab2-abcd-9c3827d97cbf" containerName="tempest-tests-tempest-tests-runner" Dec 08 20:19:37 crc kubenswrapper[4706]: E1208 20:19:37.623933 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0df0a0fe-f626-4a63-b187-678e51d1b979" containerName="collect-profiles" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.623945 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0df0a0fe-f626-4a63-b187-678e51d1b979" containerName="collect-profiles" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.624190 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0df0a0fe-f626-4a63-b187-678e51d1b979" containerName="collect-profiles" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.624305 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e673aa84-089b-4ab2-abcd-9c3827d97cbf" containerName="tempest-tests-tempest-tests-runner" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.625487 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.627123 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.628422 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tg4ng" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.720092 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.720375 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2kt8\" (UniqueName: \"kubernetes.io/projected/4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a-kube-api-access-m2kt8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.825204 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2kt8\" (UniqueName: \"kubernetes.io/projected/4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a-kube-api-access-m2kt8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.825461 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.826409 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.845804 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2kt8\" (UniqueName: \"kubernetes.io/projected/4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a-kube-api-access-m2kt8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.863547 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:37 crc kubenswrapper[4706]: I1208 20:19:37.952635 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 20:19:38 crc kubenswrapper[4706]: I1208 20:19:38.447128 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 08 20:19:38 crc kubenswrapper[4706]: I1208 20:19:38.449440 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 20:19:38 crc kubenswrapper[4706]: I1208 20:19:38.484736 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a","Type":"ContainerStarted","Data":"8e28c4de8f71cd850c995eb3312e49c06102873c2178ed28d9c1619f4b039c1c"} Dec 08 20:19:40 crc kubenswrapper[4706]: I1208 20:19:40.511151 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a","Type":"ContainerStarted","Data":"507d170d94f52bf92b59cb174314834765985604a6a4cf42ec726da0d4325c84"} Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.783457 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=26.837496982 podStartE2EDuration="27.783435675s" podCreationTimestamp="2025-12-08 20:19:37 +0000 UTC" firstStartedPulling="2025-12-08 20:19:38.449148124 +0000 UTC m=+3461.091349127" lastFinishedPulling="2025-12-08 20:19:39.395086817 +0000 UTC m=+3462.037287820" observedRunningTime="2025-12-08 20:19:40.526337759 +0000 UTC m=+3463.168538772" watchObservedRunningTime="2025-12-08 20:20:04.783435675 +0000 UTC m=+3487.425636678" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.798608 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jfnm9/must-gather-r4pxc"] Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.819849 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.834106 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jfnm9"/"default-dockercfg-fdw7d" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.834127 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jfnm9"/"openshift-service-ca.crt" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.835094 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jfnm9"/"kube-root-ca.crt" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.865984 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d1b14077-f0ed-4648-af39-7517847cbae5-must-gather-output\") pod \"must-gather-r4pxc\" (UID: \"d1b14077-f0ed-4648-af39-7517847cbae5\") " pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.866176 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx6g8\" (UniqueName: \"kubernetes.io/projected/d1b14077-f0ed-4648-af39-7517847cbae5-kube-api-access-cx6g8\") pod \"must-gather-r4pxc\" (UID: \"d1b14077-f0ed-4648-af39-7517847cbae5\") " pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.928743 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jfnm9/must-gather-r4pxc"] Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.973804 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d1b14077-f0ed-4648-af39-7517847cbae5-must-gather-output\") pod \"must-gather-r4pxc\" (UID: \"d1b14077-f0ed-4648-af39-7517847cbae5\") " pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.973884 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx6g8\" (UniqueName: \"kubernetes.io/projected/d1b14077-f0ed-4648-af39-7517847cbae5-kube-api-access-cx6g8\") pod \"must-gather-r4pxc\" (UID: \"d1b14077-f0ed-4648-af39-7517847cbae5\") " pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:20:04 crc kubenswrapper[4706]: I1208 20:20:04.974708 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d1b14077-f0ed-4648-af39-7517847cbae5-must-gather-output\") pod \"must-gather-r4pxc\" (UID: \"d1b14077-f0ed-4648-af39-7517847cbae5\") " pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:20:05 crc kubenswrapper[4706]: I1208 20:20:05.020769 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx6g8\" (UniqueName: \"kubernetes.io/projected/d1b14077-f0ed-4648-af39-7517847cbae5-kube-api-access-cx6g8\") pod \"must-gather-r4pxc\" (UID: \"d1b14077-f0ed-4648-af39-7517847cbae5\") " pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:20:05 crc kubenswrapper[4706]: I1208 20:20:05.183973 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:20:05 crc kubenswrapper[4706]: I1208 20:20:05.808909 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jfnm9/must-gather-r4pxc"] Dec 08 20:20:05 crc kubenswrapper[4706]: I1208 20:20:05.950171 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" event={"ID":"d1b14077-f0ed-4648-af39-7517847cbae5","Type":"ContainerStarted","Data":"0ec589cfebf62722d18b505af6d2feeafb13c427aebeec92807ebbbd30b48ff1"} Dec 08 20:20:14 crc kubenswrapper[4706]: I1208 20:20:14.059632 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" event={"ID":"d1b14077-f0ed-4648-af39-7517847cbae5","Type":"ContainerStarted","Data":"ba604c294f41059bfc2f36e4f782d45b0b839a7e0ca4b04f6446ed37491c226d"} Dec 08 20:20:15 crc kubenswrapper[4706]: I1208 20:20:15.082835 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" event={"ID":"d1b14077-f0ed-4648-af39-7517847cbae5","Type":"ContainerStarted","Data":"aa15767d6300fec1431c5b3c9b7258d556cf2287209048f3ed9daa84eae1873f"} Dec 08 20:20:15 crc kubenswrapper[4706]: I1208 20:20:15.107714 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" podStartSLOduration=3.472800092 podStartE2EDuration="11.107685948s" podCreationTimestamp="2025-12-08 20:20:04 +0000 UTC" firstStartedPulling="2025-12-08 20:20:05.816232142 +0000 UTC m=+3488.458433145" lastFinishedPulling="2025-12-08 20:20:13.451117998 +0000 UTC m=+3496.093319001" observedRunningTime="2025-12-08 20:20:15.104184838 +0000 UTC m=+3497.746385851" watchObservedRunningTime="2025-12-08 20:20:15.107685948 +0000 UTC m=+3497.749886951" Dec 08 20:20:18 crc kubenswrapper[4706]: I1208 20:20:18.702125 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-qw2sc"] Dec 08 20:20:18 crc kubenswrapper[4706]: I1208 20:20:18.704927 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:20:18 crc kubenswrapper[4706]: I1208 20:20:18.789769 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f3d7debc-d6be-4196-9d19-1739925b65b1-host\") pod \"crc-debug-qw2sc\" (UID: \"f3d7debc-d6be-4196-9d19-1739925b65b1\") " pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:20:18 crc kubenswrapper[4706]: I1208 20:20:18.789838 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x59lj\" (UniqueName: \"kubernetes.io/projected/f3d7debc-d6be-4196-9d19-1739925b65b1-kube-api-access-x59lj\") pod \"crc-debug-qw2sc\" (UID: \"f3d7debc-d6be-4196-9d19-1739925b65b1\") " pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:20:18 crc kubenswrapper[4706]: I1208 20:20:18.892745 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f3d7debc-d6be-4196-9d19-1739925b65b1-host\") pod \"crc-debug-qw2sc\" (UID: \"f3d7debc-d6be-4196-9d19-1739925b65b1\") " pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:20:18 crc kubenswrapper[4706]: I1208 20:20:18.893111 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x59lj\" (UniqueName: \"kubernetes.io/projected/f3d7debc-d6be-4196-9d19-1739925b65b1-kube-api-access-x59lj\") pod \"crc-debug-qw2sc\" (UID: \"f3d7debc-d6be-4196-9d19-1739925b65b1\") " pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:20:18 crc kubenswrapper[4706]: I1208 20:20:18.892951 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f3d7debc-d6be-4196-9d19-1739925b65b1-host\") pod \"crc-debug-qw2sc\" (UID: \"f3d7debc-d6be-4196-9d19-1739925b65b1\") " pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:20:18 crc kubenswrapper[4706]: I1208 20:20:18.922189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x59lj\" (UniqueName: \"kubernetes.io/projected/f3d7debc-d6be-4196-9d19-1739925b65b1-kube-api-access-x59lj\") pod \"crc-debug-qw2sc\" (UID: \"f3d7debc-d6be-4196-9d19-1739925b65b1\") " pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.033014 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.136164 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" event={"ID":"f3d7debc-d6be-4196-9d19-1739925b65b1","Type":"ContainerStarted","Data":"23bfd3a5e7c8797c3b1abaca07d5536af774857cf3a9b9222c79c0e4966c7502"} Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.466442 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xjz9d"] Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.471525 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.478332 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xjz9d"] Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.630122 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-utilities\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.630180 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvpq2\" (UniqueName: \"kubernetes.io/projected/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-kube-api-access-xvpq2\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.630215 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-catalog-content\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.733452 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-utilities\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.733522 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvpq2\" (UniqueName: \"kubernetes.io/projected/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-kube-api-access-xvpq2\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.733553 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-catalog-content\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.735462 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-utilities\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.736055 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-catalog-content\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.769337 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvpq2\" (UniqueName: \"kubernetes.io/projected/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-kube-api-access-xvpq2\") pod \"certified-operators-xjz9d\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:19 crc kubenswrapper[4706]: I1208 20:20:19.813544 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:20 crc kubenswrapper[4706]: I1208 20:20:20.628349 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xjz9d"] Dec 08 20:20:21 crc kubenswrapper[4706]: I1208 20:20:21.238744 4706 generic.go:334] "Generic (PLEG): container finished" podID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerID="2797dc87f44adc0fe14d65522eee058210a7d81a8c34d92d46e09be79ca9678a" exitCode=0 Dec 08 20:20:21 crc kubenswrapper[4706]: I1208 20:20:21.239346 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjz9d" event={"ID":"ce6bc33b-e8ac-4d35-9c58-8142b149fef4","Type":"ContainerDied","Data":"2797dc87f44adc0fe14d65522eee058210a7d81a8c34d92d46e09be79ca9678a"} Dec 08 20:20:21 crc kubenswrapper[4706]: I1208 20:20:21.239392 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjz9d" event={"ID":"ce6bc33b-e8ac-4d35-9c58-8142b149fef4","Type":"ContainerStarted","Data":"250c722ba3d0961d297edbcb6c6a9984ced6bdabcd68f2b2ac6ec559909cef53"} Dec 08 20:20:22 crc kubenswrapper[4706]: I1208 20:20:22.255995 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjz9d" event={"ID":"ce6bc33b-e8ac-4d35-9c58-8142b149fef4","Type":"ContainerStarted","Data":"2a164bc0d848d6b331f28565ad36d39eb1a291877a9a1f94def51781704b1928"} Dec 08 20:20:23 crc kubenswrapper[4706]: I1208 20:20:23.278407 4706 generic.go:334] "Generic (PLEG): container finished" podID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerID="2a164bc0d848d6b331f28565ad36d39eb1a291877a9a1f94def51781704b1928" exitCode=0 Dec 08 20:20:23 crc kubenswrapper[4706]: I1208 20:20:23.278461 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjz9d" event={"ID":"ce6bc33b-e8ac-4d35-9c58-8142b149fef4","Type":"ContainerDied","Data":"2a164bc0d848d6b331f28565ad36d39eb1a291877a9a1f94def51781704b1928"} Dec 08 20:20:24 crc kubenswrapper[4706]: I1208 20:20:24.305465 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjz9d" event={"ID":"ce6bc33b-e8ac-4d35-9c58-8142b149fef4","Type":"ContainerStarted","Data":"5aebe51770a86b1e9ac24e2f781efb5f2a1474dd29e66b5062ea416248edc876"} Dec 08 20:20:24 crc kubenswrapper[4706]: I1208 20:20:24.330760 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xjz9d" podStartSLOduration=2.792490662 podStartE2EDuration="5.33072427s" podCreationTimestamp="2025-12-08 20:20:19 +0000 UTC" firstStartedPulling="2025-12-08 20:20:21.241400354 +0000 UTC m=+3503.883601347" lastFinishedPulling="2025-12-08 20:20:23.779633952 +0000 UTC m=+3506.421834955" observedRunningTime="2025-12-08 20:20:24.326109189 +0000 UTC m=+3506.968310202" watchObservedRunningTime="2025-12-08 20:20:24.33072427 +0000 UTC m=+3506.972925273" Dec 08 20:20:29 crc kubenswrapper[4706]: I1208 20:20:29.813723 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:29 crc kubenswrapper[4706]: I1208 20:20:29.814647 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:29 crc kubenswrapper[4706]: I1208 20:20:29.888881 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:30 crc kubenswrapper[4706]: I1208 20:20:30.529604 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:30 crc kubenswrapper[4706]: I1208 20:20:30.608077 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xjz9d"] Dec 08 20:20:32 crc kubenswrapper[4706]: I1208 20:20:32.460689 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xjz9d" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerName="registry-server" containerID="cri-o://5aebe51770a86b1e9ac24e2f781efb5f2a1474dd29e66b5062ea416248edc876" gracePeriod=2 Dec 08 20:20:33 crc kubenswrapper[4706]: I1208 20:20:33.478616 4706 generic.go:334] "Generic (PLEG): container finished" podID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerID="5aebe51770a86b1e9ac24e2f781efb5f2a1474dd29e66b5062ea416248edc876" exitCode=0 Dec 08 20:20:33 crc kubenswrapper[4706]: I1208 20:20:33.478749 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjz9d" event={"ID":"ce6bc33b-e8ac-4d35-9c58-8142b149fef4","Type":"ContainerDied","Data":"5aebe51770a86b1e9ac24e2f781efb5f2a1474dd29e66b5062ea416248edc876"} Dec 08 20:20:37 crc kubenswrapper[4706]: E1208 20:20:37.092191 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Dec 08 20:20:37 crc kubenswrapper[4706]: E1208 20:20:37.093204 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x59lj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-qw2sc_openshift-must-gather-jfnm9(f3d7debc-d6be-4196-9d19-1739925b65b1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 20:20:37 crc kubenswrapper[4706]: E1208 20:20:37.094598 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" podUID="f3d7debc-d6be-4196-9d19-1739925b65b1" Dec 08 20:20:37 crc kubenswrapper[4706]: E1208 20:20:37.584442 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" podUID="f3d7debc-d6be-4196-9d19-1739925b65b1" Dec 08 20:20:37 crc kubenswrapper[4706]: I1208 20:20:37.866133 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:37 crc kubenswrapper[4706]: I1208 20:20:37.975171 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvpq2\" (UniqueName: \"kubernetes.io/projected/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-kube-api-access-xvpq2\") pod \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " Dec 08 20:20:37 crc kubenswrapper[4706]: I1208 20:20:37.975256 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-utilities\") pod \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " Dec 08 20:20:37 crc kubenswrapper[4706]: I1208 20:20:37.975381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-catalog-content\") pod \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\" (UID: \"ce6bc33b-e8ac-4d35-9c58-8142b149fef4\") " Dec 08 20:20:37 crc kubenswrapper[4706]: I1208 20:20:37.976561 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-utilities" (OuterVolumeSpecName: "utilities") pod "ce6bc33b-e8ac-4d35-9c58-8142b149fef4" (UID: "ce6bc33b-e8ac-4d35-9c58-8142b149fef4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:20:37 crc kubenswrapper[4706]: I1208 20:20:37.981871 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.003680 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-kube-api-access-xvpq2" (OuterVolumeSpecName: "kube-api-access-xvpq2") pod "ce6bc33b-e8ac-4d35-9c58-8142b149fef4" (UID: "ce6bc33b-e8ac-4d35-9c58-8142b149fef4"). InnerVolumeSpecName "kube-api-access-xvpq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.026155 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce6bc33b-e8ac-4d35-9c58-8142b149fef4" (UID: "ce6bc33b-e8ac-4d35-9c58-8142b149fef4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.084665 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvpq2\" (UniqueName: \"kubernetes.io/projected/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-kube-api-access-xvpq2\") on node \"crc\" DevicePath \"\"" Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.084731 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6bc33b-e8ac-4d35-9c58-8142b149fef4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.597675 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjz9d" event={"ID":"ce6bc33b-e8ac-4d35-9c58-8142b149fef4","Type":"ContainerDied","Data":"250c722ba3d0961d297edbcb6c6a9984ced6bdabcd68f2b2ac6ec559909cef53"} Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.597742 4706 scope.go:117] "RemoveContainer" containerID="5aebe51770a86b1e9ac24e2f781efb5f2a1474dd29e66b5062ea416248edc876" Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.598640 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjz9d" Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.646802 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xjz9d"] Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.657724 4706 scope.go:117] "RemoveContainer" containerID="2a164bc0d848d6b331f28565ad36d39eb1a291877a9a1f94def51781704b1928" Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.662776 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xjz9d"] Dec 08 20:20:38 crc kubenswrapper[4706]: I1208 20:20:38.690935 4706 scope.go:117] "RemoveContainer" containerID="2797dc87f44adc0fe14d65522eee058210a7d81a8c34d92d46e09be79ca9678a" Dec 08 20:20:39 crc kubenswrapper[4706]: I1208 20:20:39.627930 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" path="/var/lib/kubelet/pods/ce6bc33b-e8ac-4d35-9c58-8142b149fef4/volumes" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.839779 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rtglt"] Dec 08 20:20:47 crc kubenswrapper[4706]: E1208 20:20:47.841451 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerName="extract-content" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.841472 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerName="extract-content" Dec 08 20:20:47 crc kubenswrapper[4706]: E1208 20:20:47.841508 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerName="extract-utilities" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.841515 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerName="extract-utilities" Dec 08 20:20:47 crc kubenswrapper[4706]: E1208 20:20:47.841547 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerName="registry-server" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.841554 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerName="registry-server" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.841834 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce6bc33b-e8ac-4d35-9c58-8142b149fef4" containerName="registry-server" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.844278 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.881225 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtglt"] Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.949905 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbbcv\" (UniqueName: \"kubernetes.io/projected/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-kube-api-access-xbbcv\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.950114 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-utilities\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:47 crc kubenswrapper[4706]: I1208 20:20:47.950250 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-catalog-content\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:48 crc kubenswrapper[4706]: I1208 20:20:48.052768 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-catalog-content\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:48 crc kubenswrapper[4706]: I1208 20:20:48.052903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbbcv\" (UniqueName: \"kubernetes.io/projected/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-kube-api-access-xbbcv\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:48 crc kubenswrapper[4706]: I1208 20:20:48.053027 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-utilities\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:48 crc kubenswrapper[4706]: I1208 20:20:48.053508 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-catalog-content\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:48 crc kubenswrapper[4706]: I1208 20:20:48.053698 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-utilities\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:48 crc kubenswrapper[4706]: I1208 20:20:48.077624 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbbcv\" (UniqueName: \"kubernetes.io/projected/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-kube-api-access-xbbcv\") pod \"redhat-marketplace-rtglt\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:48 crc kubenswrapper[4706]: I1208 20:20:48.183871 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:48 crc kubenswrapper[4706]: I1208 20:20:48.831395 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtglt"] Dec 08 20:20:49 crc kubenswrapper[4706]: I1208 20:20:49.803417 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerID="1766c3f800f0cfe119c3f5a5284f23880fedac18ab6465cb14a7ab24aa40a9c6" exitCode=0 Dec 08 20:20:49 crc kubenswrapper[4706]: I1208 20:20:49.803530 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtglt" event={"ID":"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a","Type":"ContainerDied","Data":"1766c3f800f0cfe119c3f5a5284f23880fedac18ab6465cb14a7ab24aa40a9c6"} Dec 08 20:20:49 crc kubenswrapper[4706]: I1208 20:20:49.806091 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtglt" event={"ID":"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a","Type":"ContainerStarted","Data":"22d5d46c7b03287d5ecdf90a28dc9c97204b47546f73023fbcde786ac324243f"} Dec 08 20:20:50 crc kubenswrapper[4706]: I1208 20:20:50.825145 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtglt" event={"ID":"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a","Type":"ContainerStarted","Data":"9df5f0cd1bfbb72435795cc752d082667de32f62874ec3f03fb80db55a589b78"} Dec 08 20:20:51 crc kubenswrapper[4706]: I1208 20:20:51.838744 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" event={"ID":"f3d7debc-d6be-4196-9d19-1739925b65b1","Type":"ContainerStarted","Data":"7c42cf55b53cc00cf6cff87d096e0cba843a30019dc28fcf6fdba8bee44f3718"} Dec 08 20:20:51 crc kubenswrapper[4706]: I1208 20:20:51.843540 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerID="9df5f0cd1bfbb72435795cc752d082667de32f62874ec3f03fb80db55a589b78" exitCode=0 Dec 08 20:20:51 crc kubenswrapper[4706]: I1208 20:20:51.843605 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtglt" event={"ID":"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a","Type":"ContainerDied","Data":"9df5f0cd1bfbb72435795cc752d082667de32f62874ec3f03fb80db55a589b78"} Dec 08 20:20:51 crc kubenswrapper[4706]: I1208 20:20:51.872887 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" podStartSLOduration=1.90917725 podStartE2EDuration="33.872863047s" podCreationTimestamp="2025-12-08 20:20:18 +0000 UTC" firstStartedPulling="2025-12-08 20:20:19.082561343 +0000 UTC m=+3501.724762346" lastFinishedPulling="2025-12-08 20:20:51.04624714 +0000 UTC m=+3533.688448143" observedRunningTime="2025-12-08 20:20:51.859689283 +0000 UTC m=+3534.501890286" watchObservedRunningTime="2025-12-08 20:20:51.872863047 +0000 UTC m=+3534.515064050" Dec 08 20:20:52 crc kubenswrapper[4706]: I1208 20:20:52.861508 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtglt" event={"ID":"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a","Type":"ContainerStarted","Data":"bbd2b584a19ff57345fb0f089d6094536a9b44ed4086c66e0f38f70b467b9464"} Dec 08 20:20:52 crc kubenswrapper[4706]: I1208 20:20:52.886705 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rtglt" podStartSLOduration=3.474393701 podStartE2EDuration="5.886686476s" podCreationTimestamp="2025-12-08 20:20:47 +0000 UTC" firstStartedPulling="2025-12-08 20:20:49.806053976 +0000 UTC m=+3532.448254979" lastFinishedPulling="2025-12-08 20:20:52.218346751 +0000 UTC m=+3534.860547754" observedRunningTime="2025-12-08 20:20:52.881732415 +0000 UTC m=+3535.523933418" watchObservedRunningTime="2025-12-08 20:20:52.886686476 +0000 UTC m=+3535.528887479" Dec 08 20:20:58 crc kubenswrapper[4706]: I1208 20:20:58.184069 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:58 crc kubenswrapper[4706]: I1208 20:20:58.184809 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:58 crc kubenswrapper[4706]: I1208 20:20:58.250221 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:59 crc kubenswrapper[4706]: I1208 20:20:59.040437 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:20:59 crc kubenswrapper[4706]: I1208 20:20:59.118528 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtglt"] Dec 08 20:21:00 crc kubenswrapper[4706]: I1208 20:21:00.968615 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rtglt" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerName="registry-server" containerID="cri-o://bbd2b584a19ff57345fb0f089d6094536a9b44ed4086c66e0f38f70b467b9464" gracePeriod=2 Dec 08 20:21:01 crc kubenswrapper[4706]: I1208 20:21:01.995701 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerID="bbd2b584a19ff57345fb0f089d6094536a9b44ed4086c66e0f38f70b467b9464" exitCode=0 Dec 08 20:21:01 crc kubenswrapper[4706]: I1208 20:21:01.995773 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtglt" event={"ID":"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a","Type":"ContainerDied","Data":"bbd2b584a19ff57345fb0f089d6094536a9b44ed4086c66e0f38f70b467b9464"} Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.657510 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.769373 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbbcv\" (UniqueName: \"kubernetes.io/projected/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-kube-api-access-xbbcv\") pod \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.769546 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-utilities\") pod \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.769575 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-catalog-content\") pod \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\" (UID: \"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a\") " Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.771009 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-utilities" (OuterVolumeSpecName: "utilities") pod "d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" (UID: "d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.785339 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-kube-api-access-xbbcv" (OuterVolumeSpecName: "kube-api-access-xbbcv") pod "d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" (UID: "d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a"). InnerVolumeSpecName "kube-api-access-xbbcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.792960 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" (UID: "d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.872418 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbbcv\" (UniqueName: \"kubernetes.io/projected/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-kube-api-access-xbbcv\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.872466 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:07 crc kubenswrapper[4706]: I1208 20:21:07.872480 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:08 crc kubenswrapper[4706]: I1208 20:21:08.075820 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtglt" event={"ID":"d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a","Type":"ContainerDied","Data":"22d5d46c7b03287d5ecdf90a28dc9c97204b47546f73023fbcde786ac324243f"} Dec 08 20:21:08 crc kubenswrapper[4706]: I1208 20:21:08.076202 4706 scope.go:117] "RemoveContainer" containerID="bbd2b584a19ff57345fb0f089d6094536a9b44ed4086c66e0f38f70b467b9464" Dec 08 20:21:08 crc kubenswrapper[4706]: I1208 20:21:08.076012 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtglt" Dec 08 20:21:08 crc kubenswrapper[4706]: I1208 20:21:08.122534 4706 scope.go:117] "RemoveContainer" containerID="9df5f0cd1bfbb72435795cc752d082667de32f62874ec3f03fb80db55a589b78" Dec 08 20:21:08 crc kubenswrapper[4706]: I1208 20:21:08.127966 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtglt"] Dec 08 20:21:08 crc kubenswrapper[4706]: I1208 20:21:08.140553 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtglt"] Dec 08 20:21:08 crc kubenswrapper[4706]: I1208 20:21:08.155517 4706 scope.go:117] "RemoveContainer" containerID="1766c3f800f0cfe119c3f5a5284f23880fedac18ab6465cb14a7ab24aa40a9c6" Dec 08 20:21:09 crc kubenswrapper[4706]: I1208 20:21:09.622744 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" path="/var/lib/kubelet/pods/d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a/volumes" Dec 08 20:21:35 crc kubenswrapper[4706]: I1208 20:21:35.836206 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:21:35 crc kubenswrapper[4706]: I1208 20:21:35.837029 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:21:44 crc kubenswrapper[4706]: I1208 20:21:44.520200 4706 generic.go:334] "Generic (PLEG): container finished" podID="f3d7debc-d6be-4196-9d19-1739925b65b1" containerID="7c42cf55b53cc00cf6cff87d096e0cba843a30019dc28fcf6fdba8bee44f3718" exitCode=0 Dec 08 20:21:44 crc kubenswrapper[4706]: I1208 20:21:44.520315 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" event={"ID":"f3d7debc-d6be-4196-9d19-1739925b65b1","Type":"ContainerDied","Data":"7c42cf55b53cc00cf6cff87d096e0cba843a30019dc28fcf6fdba8bee44f3718"} Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.670302 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.720884 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-qw2sc"] Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.733191 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-qw2sc"] Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.735507 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x59lj\" (UniqueName: \"kubernetes.io/projected/f3d7debc-d6be-4196-9d19-1739925b65b1-kube-api-access-x59lj\") pod \"f3d7debc-d6be-4196-9d19-1739925b65b1\" (UID: \"f3d7debc-d6be-4196-9d19-1739925b65b1\") " Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.735868 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f3d7debc-d6be-4196-9d19-1739925b65b1-host\") pod \"f3d7debc-d6be-4196-9d19-1739925b65b1\" (UID: \"f3d7debc-d6be-4196-9d19-1739925b65b1\") " Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.736195 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f3d7debc-d6be-4196-9d19-1739925b65b1-host" (OuterVolumeSpecName: "host") pod "f3d7debc-d6be-4196-9d19-1739925b65b1" (UID: "f3d7debc-d6be-4196-9d19-1739925b65b1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.737858 4706 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f3d7debc-d6be-4196-9d19-1739925b65b1-host\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.745137 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3d7debc-d6be-4196-9d19-1739925b65b1-kube-api-access-x59lj" (OuterVolumeSpecName: "kube-api-access-x59lj") pod "f3d7debc-d6be-4196-9d19-1739925b65b1" (UID: "f3d7debc-d6be-4196-9d19-1739925b65b1"). InnerVolumeSpecName "kube-api-access-x59lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:21:45 crc kubenswrapper[4706]: I1208 20:21:45.840681 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x59lj\" (UniqueName: \"kubernetes.io/projected/f3d7debc-d6be-4196-9d19-1739925b65b1-kube-api-access-x59lj\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.556535 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23bfd3a5e7c8797c3b1abaca07d5536af774857cf3a9b9222c79c0e4966c7502" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.556619 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-qw2sc" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.904313 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-k8rsx"] Dec 08 20:21:46 crc kubenswrapper[4706]: E1208 20:21:46.906056 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerName="registry-server" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.906233 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerName="registry-server" Dec 08 20:21:46 crc kubenswrapper[4706]: E1208 20:21:46.906371 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerName="extract-content" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.906439 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerName="extract-content" Dec 08 20:21:46 crc kubenswrapper[4706]: E1208 20:21:46.906513 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerName="extract-utilities" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.906569 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerName="extract-utilities" Dec 08 20:21:46 crc kubenswrapper[4706]: E1208 20:21:46.906636 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d7debc-d6be-4196-9d19-1739925b65b1" containerName="container-00" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.906696 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d7debc-d6be-4196-9d19-1739925b65b1" containerName="container-00" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.907005 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d7debc-d6be-4196-9d19-1739925b65b1" containerName="container-00" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.907086 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ca9245-0ecd-47e2-a5a8-4ca9247ec86a" containerName="registry-server" Dec 08 20:21:46 crc kubenswrapper[4706]: I1208 20:21:46.908136 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.074019 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvdnr\" (UniqueName: \"kubernetes.io/projected/9f4c1920-d348-4880-938b-a278d98468ea-kube-api-access-tvdnr\") pod \"crc-debug-k8rsx\" (UID: \"9f4c1920-d348-4880-938b-a278d98468ea\") " pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.074082 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f4c1920-d348-4880-938b-a278d98468ea-host\") pod \"crc-debug-k8rsx\" (UID: \"9f4c1920-d348-4880-938b-a278d98468ea\") " pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.176377 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvdnr\" (UniqueName: \"kubernetes.io/projected/9f4c1920-d348-4880-938b-a278d98468ea-kube-api-access-tvdnr\") pod \"crc-debug-k8rsx\" (UID: \"9f4c1920-d348-4880-938b-a278d98468ea\") " pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.176474 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f4c1920-d348-4880-938b-a278d98468ea-host\") pod \"crc-debug-k8rsx\" (UID: \"9f4c1920-d348-4880-938b-a278d98468ea\") " pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.176759 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f4c1920-d348-4880-938b-a278d98468ea-host\") pod \"crc-debug-k8rsx\" (UID: \"9f4c1920-d348-4880-938b-a278d98468ea\") " pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.201128 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvdnr\" (UniqueName: \"kubernetes.io/projected/9f4c1920-d348-4880-938b-a278d98468ea-kube-api-access-tvdnr\") pod \"crc-debug-k8rsx\" (UID: \"9f4c1920-d348-4880-938b-a278d98468ea\") " pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.273993 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.569964 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" event={"ID":"9f4c1920-d348-4880-938b-a278d98468ea","Type":"ContainerStarted","Data":"783eb9cee8bfac79608920a1c7a9351f209f6311c6cf39b2254b7339915ae86e"} Dec 08 20:21:47 crc kubenswrapper[4706]: I1208 20:21:47.625955 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3d7debc-d6be-4196-9d19-1739925b65b1" path="/var/lib/kubelet/pods/f3d7debc-d6be-4196-9d19-1739925b65b1/volumes" Dec 08 20:21:48 crc kubenswrapper[4706]: I1208 20:21:48.584116 4706 generic.go:334] "Generic (PLEG): container finished" podID="9f4c1920-d348-4880-938b-a278d98468ea" containerID="f31f54de62a095892a4a75e8139f828219b3f2f55d50859ce7900c7b01a85bc3" exitCode=0 Dec 08 20:21:48 crc kubenswrapper[4706]: I1208 20:21:48.584202 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" event={"ID":"9f4c1920-d348-4880-938b-a278d98468ea","Type":"ContainerDied","Data":"f31f54de62a095892a4a75e8139f828219b3f2f55d50859ce7900c7b01a85bc3"} Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.759940 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.796253 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-k8rsx"] Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.813068 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-k8rsx"] Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.851462 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f4c1920-d348-4880-938b-a278d98468ea-host\") pod \"9f4c1920-d348-4880-938b-a278d98468ea\" (UID: \"9f4c1920-d348-4880-938b-a278d98468ea\") " Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.851628 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f4c1920-d348-4880-938b-a278d98468ea-host" (OuterVolumeSpecName: "host") pod "9f4c1920-d348-4880-938b-a278d98468ea" (UID: "9f4c1920-d348-4880-938b-a278d98468ea"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.851869 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvdnr\" (UniqueName: \"kubernetes.io/projected/9f4c1920-d348-4880-938b-a278d98468ea-kube-api-access-tvdnr\") pod \"9f4c1920-d348-4880-938b-a278d98468ea\" (UID: \"9f4c1920-d348-4880-938b-a278d98468ea\") " Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.852996 4706 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f4c1920-d348-4880-938b-a278d98468ea-host\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.860597 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f4c1920-d348-4880-938b-a278d98468ea-kube-api-access-tvdnr" (OuterVolumeSpecName: "kube-api-access-tvdnr") pod "9f4c1920-d348-4880-938b-a278d98468ea" (UID: "9f4c1920-d348-4880-938b-a278d98468ea"). InnerVolumeSpecName "kube-api-access-tvdnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:21:49 crc kubenswrapper[4706]: I1208 20:21:49.955411 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvdnr\" (UniqueName: \"kubernetes.io/projected/9f4c1920-d348-4880-938b-a278d98468ea-kube-api-access-tvdnr\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:50 crc kubenswrapper[4706]: I1208 20:21:50.609273 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="783eb9cee8bfac79608920a1c7a9351f209f6311c6cf39b2254b7339915ae86e" Dec 08 20:21:50 crc kubenswrapper[4706]: I1208 20:21:50.609582 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-k8rsx" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.002156 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-f7bg6"] Dec 08 20:21:51 crc kubenswrapper[4706]: E1208 20:21:51.002678 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4c1920-d348-4880-938b-a278d98468ea" containerName="container-00" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.002693 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4c1920-d348-4880-938b-a278d98468ea" containerName="container-00" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.002899 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f4c1920-d348-4880-938b-a278d98468ea" containerName="container-00" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.003726 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.084689 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssxf2\" (UniqueName: \"kubernetes.io/projected/b1740adf-239d-48a9-bc00-640421bf8ee2-kube-api-access-ssxf2\") pod \"crc-debug-f7bg6\" (UID: \"b1740adf-239d-48a9-bc00-640421bf8ee2\") " pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.085554 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1740adf-239d-48a9-bc00-640421bf8ee2-host\") pod \"crc-debug-f7bg6\" (UID: \"b1740adf-239d-48a9-bc00-640421bf8ee2\") " pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.188134 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1740adf-239d-48a9-bc00-640421bf8ee2-host\") pod \"crc-debug-f7bg6\" (UID: \"b1740adf-239d-48a9-bc00-640421bf8ee2\") " pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.188842 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssxf2\" (UniqueName: \"kubernetes.io/projected/b1740adf-239d-48a9-bc00-640421bf8ee2-kube-api-access-ssxf2\") pod \"crc-debug-f7bg6\" (UID: \"b1740adf-239d-48a9-bc00-640421bf8ee2\") " pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.188486 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1740adf-239d-48a9-bc00-640421bf8ee2-host\") pod \"crc-debug-f7bg6\" (UID: \"b1740adf-239d-48a9-bc00-640421bf8ee2\") " pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.211617 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssxf2\" (UniqueName: \"kubernetes.io/projected/b1740adf-239d-48a9-bc00-640421bf8ee2-kube-api-access-ssxf2\") pod \"crc-debug-f7bg6\" (UID: \"b1740adf-239d-48a9-bc00-640421bf8ee2\") " pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.326667 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:51 crc kubenswrapper[4706]: W1208 20:21:51.362778 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1740adf_239d_48a9_bc00_640421bf8ee2.slice/crio-78dd2bbea33b02e4fcf25f22263764e2897d21b68bcd150f073cc189f45e17d3 WatchSource:0}: Error finding container 78dd2bbea33b02e4fcf25f22263764e2897d21b68bcd150f073cc189f45e17d3: Status 404 returned error can't find the container with id 78dd2bbea33b02e4fcf25f22263764e2897d21b68bcd150f073cc189f45e17d3 Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.646934 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f4c1920-d348-4880-938b-a278d98468ea" path="/var/lib/kubelet/pods/9f4c1920-d348-4880-938b-a278d98468ea/volumes" Dec 08 20:21:51 crc kubenswrapper[4706]: I1208 20:21:51.651679 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" event={"ID":"b1740adf-239d-48a9-bc00-640421bf8ee2","Type":"ContainerStarted","Data":"78dd2bbea33b02e4fcf25f22263764e2897d21b68bcd150f073cc189f45e17d3"} Dec 08 20:21:52 crc kubenswrapper[4706]: I1208 20:21:52.661933 4706 generic.go:334] "Generic (PLEG): container finished" podID="b1740adf-239d-48a9-bc00-640421bf8ee2" containerID="9d08be77879f704e535d3a6f1f06700600c03a281197f71a7e7fe797a6e9bbf3" exitCode=0 Dec 08 20:21:52 crc kubenswrapper[4706]: I1208 20:21:52.662036 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" event={"ID":"b1740adf-239d-48a9-bc00-640421bf8ee2","Type":"ContainerDied","Data":"9d08be77879f704e535d3a6f1f06700600c03a281197f71a7e7fe797a6e9bbf3"} Dec 08 20:21:52 crc kubenswrapper[4706]: I1208 20:21:52.722205 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-f7bg6"] Dec 08 20:21:52 crc kubenswrapper[4706]: I1208 20:21:52.734082 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jfnm9/crc-debug-f7bg6"] Dec 08 20:21:53 crc kubenswrapper[4706]: I1208 20:21:53.850579 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:53 crc kubenswrapper[4706]: I1208 20:21:53.962908 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1740adf-239d-48a9-bc00-640421bf8ee2-host\") pod \"b1740adf-239d-48a9-bc00-640421bf8ee2\" (UID: \"b1740adf-239d-48a9-bc00-640421bf8ee2\") " Dec 08 20:21:53 crc kubenswrapper[4706]: I1208 20:21:53.963101 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1740adf-239d-48a9-bc00-640421bf8ee2-host" (OuterVolumeSpecName: "host") pod "b1740adf-239d-48a9-bc00-640421bf8ee2" (UID: "b1740adf-239d-48a9-bc00-640421bf8ee2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 20:21:53 crc kubenswrapper[4706]: I1208 20:21:53.963333 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssxf2\" (UniqueName: \"kubernetes.io/projected/b1740adf-239d-48a9-bc00-640421bf8ee2-kube-api-access-ssxf2\") pod \"b1740adf-239d-48a9-bc00-640421bf8ee2\" (UID: \"b1740adf-239d-48a9-bc00-640421bf8ee2\") " Dec 08 20:21:53 crc kubenswrapper[4706]: I1208 20:21:53.963985 4706 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1740adf-239d-48a9-bc00-640421bf8ee2-host\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:53 crc kubenswrapper[4706]: I1208 20:21:53.975679 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1740adf-239d-48a9-bc00-640421bf8ee2-kube-api-access-ssxf2" (OuterVolumeSpecName: "kube-api-access-ssxf2") pod "b1740adf-239d-48a9-bc00-640421bf8ee2" (UID: "b1740adf-239d-48a9-bc00-640421bf8ee2"). InnerVolumeSpecName "kube-api-access-ssxf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:21:54 crc kubenswrapper[4706]: I1208 20:21:54.066076 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssxf2\" (UniqueName: \"kubernetes.io/projected/b1740adf-239d-48a9-bc00-640421bf8ee2-kube-api-access-ssxf2\") on node \"crc\" DevicePath \"\"" Dec 08 20:21:54 crc kubenswrapper[4706]: I1208 20:21:54.701326 4706 scope.go:117] "RemoveContainer" containerID="9d08be77879f704e535d3a6f1f06700600c03a281197f71a7e7fe797a6e9bbf3" Dec 08 20:21:54 crc kubenswrapper[4706]: I1208 20:21:54.701366 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/crc-debug-f7bg6" Dec 08 20:21:55 crc kubenswrapper[4706]: I1208 20:21:55.639706 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1740adf-239d-48a9-bc00-640421bf8ee2" path="/var/lib/kubelet/pods/b1740adf-239d-48a9-bc00-640421bf8ee2/volumes" Dec 08 20:22:05 crc kubenswrapper[4706]: I1208 20:22:05.836906 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:22:05 crc kubenswrapper[4706]: I1208 20:22:05.837832 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:22:19 crc kubenswrapper[4706]: I1208 20:22:19.910108 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88233d80-672d-47c3-9a26-70c22f748e29/init-config-reloader/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.141833 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88233d80-672d-47c3-9a26-70c22f748e29/init-config-reloader/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.170378 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88233d80-672d-47c3-9a26-70c22f748e29/alertmanager/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.174961 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88233d80-672d-47c3-9a26-70c22f748e29/config-reloader/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.385449 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8466877bdb-zbttb_b0acc09f-0610-4121-b42b-ebe7274b5c5d/barbican-api/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.394725 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8466877bdb-zbttb_b0acc09f-0610-4121-b42b-ebe7274b5c5d/barbican-api-log/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.436358 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-788468f6bb-gm7rb_aa3a5522-ce55-49b7-a575-3f83c373b2ac/barbican-keystone-listener/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.703605 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b989c8b75-k95kd_2cc4f4da-55de-4f4e-9e3e-099e148987a2/barbican-worker/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.783095 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-788468f6bb-gm7rb_aa3a5522-ce55-49b7-a575-3f83c373b2ac/barbican-keystone-listener-log/0.log" Dec 08 20:22:20 crc kubenswrapper[4706]: I1208 20:22:20.805097 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b989c8b75-k95kd_2cc4f4da-55de-4f4e-9e3e-099e148987a2/barbican-worker-log/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.088306 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-267vl_91cd912d-3492-4709-99e9-94fa496bde14/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.121712 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42/ceilometer-central-agent/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.279401 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42/proxy-httpd/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.296040 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42/sg-core/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.354415 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42/ceilometer-notification-agent/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.552362 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_81a640d0-b89a-49be-aa17-00ad9d0d444a/cinder-api/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.580662 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_81a640d0-b89a-49be-aa17-00ad9d0d444a/cinder-api-log/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.777696 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_33d60263-f73a-470d-b0de-60eb2a8ecfd5/cinder-scheduler/0.log" Dec 08 20:22:21 crc kubenswrapper[4706]: I1208 20:22:21.938547 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_33d60263-f73a-470d-b0de-60eb2a8ecfd5/probe/0.log" Dec 08 20:22:22 crc kubenswrapper[4706]: I1208 20:22:22.078889 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_b0316b44-87fb-44dd-890c-dace6fb0ab3e/cloudkitty-api/0.log" Dec 08 20:22:22 crc kubenswrapper[4706]: I1208 20:22:22.097472 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_b0316b44-87fb-44dd-890c-dace6fb0ab3e/cloudkitty-api-log/0.log" Dec 08 20:22:22 crc kubenswrapper[4706]: I1208 20:22:22.257201 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_a7e3aaf0-5179-4837-b6ca-53276d637d10/loki-compactor/0.log" Dec 08 20:22:22 crc kubenswrapper[4706]: I1208 20:22:22.350105 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-664b687b54-zfbc5_081f2271-cba5-43e4-a179-7fd4389c6531/loki-distributor/0.log" Dec 08 20:22:22 crc kubenswrapper[4706]: I1208 20:22:22.517672 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-bc75944f-7tk8c_d9d23ed1-9384-4bf8-adeb-2bc3340279ed/gateway/0.log" Dec 08 20:22:22 crc kubenswrapper[4706]: I1208 20:22:22.665688 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-bc75944f-m4nfc_7811fdf0-d557-47cb-b700-4a072602f8c7/gateway/0.log" Dec 08 20:22:22 crc kubenswrapper[4706]: I1208 20:22:22.813249 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_4d1845b4-bbc3-4661-83d4-72b972fa288a/loki-index-gateway/0.log" Dec 08 20:22:23 crc kubenswrapper[4706]: I1208 20:22:23.422290 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_9a912025-8f8e-4d61-861e-c933dafe4660/loki-ingester/0.log" Dec 08 20:22:23 crc kubenswrapper[4706]: I1208 20:22:23.430670 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt_fbd5715c-12b1-4660-bdd6-39087b0ae32a/loki-query-frontend/0.log" Dec 08 20:22:23 crc kubenswrapper[4706]: I1208 20:22:23.959817 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx_784135f7-ff54-43df-b7e7-ea4f0b940fdc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:24 crc kubenswrapper[4706]: I1208 20:22:24.336570 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-5467947bf7-dxbfm_a9df4fe8-4ed5-4bd9-a514-9ed4259283b1/loki-querier/0.log" Dec 08 20:22:24 crc kubenswrapper[4706]: I1208 20:22:24.339718 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-swwvh_9576d118-0baa-462a-b75a-a31af5b84c66/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:24 crc kubenswrapper[4706]: I1208 20:22:24.426871 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-862hf_797f5d64-d90a-4466-95f7-a9b4b43d9f1c/init/0.log" Dec 08 20:22:24 crc kubenswrapper[4706]: I1208 20:22:24.643959 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-862hf_797f5d64-d90a-4466-95f7-a9b4b43d9f1c/init/0.log" Dec 08 20:22:24 crc kubenswrapper[4706]: I1208 20:22:24.671601 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-862hf_797f5d64-d90a-4466-95f7-a9b4b43d9f1c/dnsmasq-dns/0.log" Dec 08 20:22:24 crc kubenswrapper[4706]: I1208 20:22:24.730361 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s_a9260164-eeea-47a2-9502-1744cba85586/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:24 crc kubenswrapper[4706]: I1208 20:22:24.992413 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8424f86e-b640-44f6-abd9-dc8c7a5a55cc/glance-httpd/0.log" Dec 08 20:22:25 crc kubenswrapper[4706]: I1208 20:22:25.006346 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8424f86e-b640-44f6-abd9-dc8c7a5a55cc/glance-log/0.log" Dec 08 20:22:25 crc kubenswrapper[4706]: I1208 20:22:25.399219 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6249f0fa-1385-4546-89c3-60a91e24da43/glance-httpd/0.log" Dec 08 20:22:25 crc kubenswrapper[4706]: I1208 20:22:25.403274 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6249f0fa-1385-4546-89c3-60a91e24da43/glance-log/0.log" Dec 08 20:22:25 crc kubenswrapper[4706]: I1208 20:22:25.468758 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4_42663da5-6fa0-4332-8222-63fd50d7484e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:25 crc kubenswrapper[4706]: I1208 20:22:25.687085 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-zjn26_e30a59fd-9346-4fae-8a9d-14c8700fd608/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:26 crc kubenswrapper[4706]: I1208 20:22:26.052441 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29420401-b95cm_68bc049a-f060-4de4-8be5-c88c55abebd9/keystone-cron/0.log" Dec 08 20:22:26 crc kubenswrapper[4706]: I1208 20:22:26.354141 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-786cfc44fd-zdf7t_93159ff6-eda3-42c1-94db-46a5d0777a47/keystone-api/0.log" Dec 08 20:22:26 crc kubenswrapper[4706]: I1208 20:22:26.406545 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_aa4c62bf-4f94-4e1e-afd5-9036da150cc4/kube-state-metrics/0.log" Dec 08 20:22:26 crc kubenswrapper[4706]: I1208 20:22:26.531527 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn_93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:27 crc kubenswrapper[4706]: I1208 20:22:27.079919 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79bc9d977-gw6pk_94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43/neutron-api/0.log" Dec 08 20:22:27 crc kubenswrapper[4706]: I1208 20:22:27.122180 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79bc9d977-gw6pk_94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43/neutron-httpd/0.log" Dec 08 20:22:27 crc kubenswrapper[4706]: I1208 20:22:27.380194 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb_543c86ef-f2f1-4b75-9de2-d613819b3e9a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:28 crc kubenswrapper[4706]: I1208 20:22:28.101930 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_48240f0b-4bff-4e59-b151-7e2922ee95f0/nova-api-log/0.log" Dec 08 20:22:28 crc kubenswrapper[4706]: I1208 20:22:28.230614 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_48240f0b-4bff-4e59-b151-7e2922ee95f0/nova-api-api/0.log" Dec 08 20:22:28 crc kubenswrapper[4706]: I1208 20:22:28.278737 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_667cc0c8-9292-4cbd-a013-bd645c4b57d7/nova-cell0-conductor-conductor/0.log" Dec 08 20:22:28 crc kubenswrapper[4706]: I1208 20:22:28.766200 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_734cc0d6-b154-4980-afbe-67d4d1b9b675/nova-cell1-conductor-conductor/0.log" Dec 08 20:22:28 crc kubenswrapper[4706]: I1208 20:22:28.830159 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8d07dfb3-a2ed-4e66-80a3-3288ff3f8669/nova-cell1-novncproxy-novncproxy/0.log" Dec 08 20:22:28 crc kubenswrapper[4706]: I1208 20:22:28.844172 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_ad4728d1-aa2e-4ab5-81dc-22f56a79e874/cloudkitty-proc/0.log" Dec 08 20:22:29 crc kubenswrapper[4706]: I1208 20:22:29.116413 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-m2jvl_43b41201-b854-4025-b70b-4455fcd8fc3c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:29 crc kubenswrapper[4706]: I1208 20:22:29.238605 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7eb161ee-d8ee-45e4-9fce-a0661293603a/nova-metadata-log/0.log" Dec 08 20:22:29 crc kubenswrapper[4706]: I1208 20:22:29.805945 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a0ae432b-2caa-458f-9ac4-5172e21ef898/nova-scheduler-scheduler/0.log" Dec 08 20:22:30 crc kubenswrapper[4706]: I1208 20:22:30.028119 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_13532f07-b125-4d67-bdb8-006b583bea40/mysql-bootstrap/0.log" Dec 08 20:22:30 crc kubenswrapper[4706]: I1208 20:22:30.234758 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_13532f07-b125-4d67-bdb8-006b583bea40/mysql-bootstrap/0.log" Dec 08 20:22:30 crc kubenswrapper[4706]: I1208 20:22:30.294682 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_13532f07-b125-4d67-bdb8-006b583bea40/galera/0.log" Dec 08 20:22:30 crc kubenswrapper[4706]: I1208 20:22:30.523496 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_156710b7-7dbe-4bb7-9c69-ce726d4b7a4c/mysql-bootstrap/0.log" Dec 08 20:22:30 crc kubenswrapper[4706]: I1208 20:22:30.741154 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_156710b7-7dbe-4bb7-9c69-ce726d4b7a4c/galera/0.log" Dec 08 20:22:30 crc kubenswrapper[4706]: I1208 20:22:30.775682 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_156710b7-7dbe-4bb7-9c69-ce726d4b7a4c/mysql-bootstrap/0.log" Dec 08 20:22:30 crc kubenswrapper[4706]: I1208 20:22:30.931855 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7eb161ee-d8ee-45e4-9fce-a0661293603a/nova-metadata-metadata/0.log" Dec 08 20:22:30 crc kubenswrapper[4706]: I1208 20:22:30.995087 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1b714f3f-e93c-4a5b-ae82-1f551b163fe6/openstackclient/0.log" Dec 08 20:22:31 crc kubenswrapper[4706]: I1208 20:22:31.155836 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-78wps_79f551cd-d08d-4739-9041-4543f05fbb60/ovn-controller/0.log" Dec 08 20:22:31 crc kubenswrapper[4706]: I1208 20:22:31.296738 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-schr5_a00cac1b-92f7-49a1-bbdc-97c15777b09a/openstack-network-exporter/0.log" Dec 08 20:22:31 crc kubenswrapper[4706]: I1208 20:22:31.511143 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb6gx_25287a9f-27c5-47a7-945e-684eea0157c8/ovsdb-server-init/0.log" Dec 08 20:22:31 crc kubenswrapper[4706]: I1208 20:22:31.802609 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb6gx_25287a9f-27c5-47a7-945e-684eea0157c8/ovsdb-server/0.log" Dec 08 20:22:31 crc kubenswrapper[4706]: I1208 20:22:31.840461 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb6gx_25287a9f-27c5-47a7-945e-684eea0157c8/ovsdb-server-init/0.log" Dec 08 20:22:31 crc kubenswrapper[4706]: I1208 20:22:31.869439 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb6gx_25287a9f-27c5-47a7-945e-684eea0157c8/ovs-vswitchd/0.log" Dec 08 20:22:32 crc kubenswrapper[4706]: I1208 20:22:32.153825 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-zxsvx_2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:32 crc kubenswrapper[4706]: I1208 20:22:32.207628 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_240f9d54-934b-4d65-976e-a0a5a9ca7908/openstack-network-exporter/0.log" Dec 08 20:22:32 crc kubenswrapper[4706]: I1208 20:22:32.414598 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_240f9d54-934b-4d65-976e-a0a5a9ca7908/ovn-northd/0.log" Dec 08 20:22:32 crc kubenswrapper[4706]: I1208 20:22:32.565828 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f498bda7-c379-4808-bd6c-8e5513c4322d/openstack-network-exporter/0.log" Dec 08 20:22:32 crc kubenswrapper[4706]: I1208 20:22:32.627076 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f498bda7-c379-4808-bd6c-8e5513c4322d/ovsdbserver-nb/0.log" Dec 08 20:22:32 crc kubenswrapper[4706]: I1208 20:22:32.811973 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_40f5d973-ac12-469a-a1d9-aa871b7cc9a3/ovsdbserver-sb/0.log" Dec 08 20:22:32 crc kubenswrapper[4706]: I1208 20:22:32.869311 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_40f5d973-ac12-469a-a1d9-aa871b7cc9a3/openstack-network-exporter/0.log" Dec 08 20:22:33 crc kubenswrapper[4706]: I1208 20:22:33.088886 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7494d94d4d-hn8rd_4abc8b62-e549-41c9-8bc9-8d35f3074e79/placement-api/0.log" Dec 08 20:22:33 crc kubenswrapper[4706]: I1208 20:22:33.249985 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7494d94d4d-hn8rd_4abc8b62-e549-41c9-8bc9-8d35f3074e79/placement-log/0.log" Dec 08 20:22:33 crc kubenswrapper[4706]: I1208 20:22:33.321121 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/init-config-reloader/0.log" Dec 08 20:22:33 crc kubenswrapper[4706]: I1208 20:22:33.727294 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/init-config-reloader/0.log" Dec 08 20:22:33 crc kubenswrapper[4706]: I1208 20:22:33.739438 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/thanos-sidecar/0.log" Dec 08 20:22:33 crc kubenswrapper[4706]: I1208 20:22:33.780304 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/prometheus/0.log" Dec 08 20:22:33 crc kubenswrapper[4706]: I1208 20:22:33.782837 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/config-reloader/0.log" Dec 08 20:22:33 crc kubenswrapper[4706]: I1208 20:22:33.993285 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ecc9ad61-4d9c-4d26-a240-398cfc8c3165/setup-container/0.log" Dec 08 20:22:34 crc kubenswrapper[4706]: I1208 20:22:34.335206 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ecc9ad61-4d9c-4d26-a240-398cfc8c3165/setup-container/0.log" Dec 08 20:22:34 crc kubenswrapper[4706]: I1208 20:22:34.436534 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ecc9ad61-4d9c-4d26-a240-398cfc8c3165/rabbitmq/0.log" Dec 08 20:22:34 crc kubenswrapper[4706]: I1208 20:22:34.450718 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_faed709c-224b-4c30-9af3-d64d34525599/setup-container/0.log" Dec 08 20:22:34 crc kubenswrapper[4706]: I1208 20:22:34.761317 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_faed709c-224b-4c30-9af3-d64d34525599/setup-container/0.log" Dec 08 20:22:34 crc kubenswrapper[4706]: I1208 20:22:34.948174 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_faed709c-224b-4c30-9af3-d64d34525599/rabbitmq/0.log" Dec 08 20:22:34 crc kubenswrapper[4706]: I1208 20:22:34.963575 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f_5d9ea4d5-f428-449d-baf9-880c8a2637aa/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.201759 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6_ad5354c5-7ee5-4e8a-8e20-84ac36e97e75/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.284833 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-hvs9v_074561a7-7190-4458-a9aa-1dac832175fb/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.580146 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jkcr5_893e5057-db47-44b6-b3da-f4e59a5992c5/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.704940 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-rhjj4_adf18455-0f5d-4d8e-a46d-c77523de9f84/ssh-known-hosts-edpm-deployment/0.log" Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.837052 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.837135 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.837200 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.838349 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae3b6b90e663cc0e30fdffcaf7a2453642e4c1963568b4eafa4406e240816b79"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 20:22:35 crc kubenswrapper[4706]: I1208 20:22:35.838415 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://ae3b6b90e663cc0e30fdffcaf7a2453642e4c1963568b4eafa4406e240816b79" gracePeriod=600 Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.026440 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d86ddd67f-5b6vx_637d7d72-7ba8-4a35-b37c-35aef7813c75/proxy-httpd/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.039970 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d86ddd67f-5b6vx_637d7d72-7ba8-4a35-b37c-35aef7813c75/proxy-server/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.212817 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-thlcc_a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c/swift-ring-rebalance/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.272432 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="ae3b6b90e663cc0e30fdffcaf7a2453642e4c1963568b4eafa4406e240816b79" exitCode=0 Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.272512 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"ae3b6b90e663cc0e30fdffcaf7a2453642e4c1963568b4eafa4406e240816b79"} Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.272560 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451"} Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.272584 4706 scope.go:117] "RemoveContainer" containerID="377e6bbabf0b7beac8280f9ef8a871b5e612bb5d42d0de1c2fd006d1811031b5" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.334678 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/account-auditor/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.468114 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/account-reaper/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.548284 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/account-replicator/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.708867 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/account-server/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.743644 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/container-auditor/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.771890 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/container-replicator/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.882868 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/container-server/0.log" Dec 08 20:22:36 crc kubenswrapper[4706]: I1208 20:22:36.988933 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/container-updater/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.108039 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-expirer/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.114740 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-auditor/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.213079 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-replicator/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.274541 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-server/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.346912 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/rsync/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.366018 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-updater/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.473418 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/swift-recon-cron/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.669085 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p_39dc4f79-e40f-479c-b82b-8ac18056c9ec/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.821800 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_e673aa84-089b-4ab2-abcd-9c3827d97cbf/tempest-tests-tempest-tests-runner/0.log" Dec 08 20:22:37 crc kubenswrapper[4706]: I1208 20:22:37.949241 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a/test-operator-logs-container/0.log" Dec 08 20:22:38 crc kubenswrapper[4706]: I1208 20:22:38.133240 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s_7ea46b46-9aa7-46f3-a235-8eb100586d2f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:22:40 crc kubenswrapper[4706]: I1208 20:22:40.840183 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fd927fec-2b54-41d2-affb-19f9fd8c96ab/memcached/0.log" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.632722 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-htq42"] Dec 08 20:22:41 crc kubenswrapper[4706]: E1208 20:22:41.633794 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1740adf-239d-48a9-bc00-640421bf8ee2" containerName="container-00" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.633817 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1740adf-239d-48a9-bc00-640421bf8ee2" containerName="container-00" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.634078 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1740adf-239d-48a9-bc00-640421bf8ee2" containerName="container-00" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.636221 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.670992 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-htq42"] Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.684304 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-utilities\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.684502 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rlhz\" (UniqueName: \"kubernetes.io/projected/c23056d6-d516-4856-9309-e0523fd252d4-kube-api-access-4rlhz\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.684596 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-catalog-content\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.787971 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rlhz\" (UniqueName: \"kubernetes.io/projected/c23056d6-d516-4856-9309-e0523fd252d4-kube-api-access-4rlhz\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.788080 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-catalog-content\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.788133 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-utilities\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.788759 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-utilities\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.788966 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-catalog-content\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:41 crc kubenswrapper[4706]: I1208 20:22:41.814997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rlhz\" (UniqueName: \"kubernetes.io/projected/c23056d6-d516-4856-9309-e0523fd252d4-kube-api-access-4rlhz\") pod \"community-operators-htq42\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:42 crc kubenswrapper[4706]: I1208 20:22:42.012396 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:42 crc kubenswrapper[4706]: I1208 20:22:42.730975 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-htq42"] Dec 08 20:22:43 crc kubenswrapper[4706]: I1208 20:22:43.426066 4706 generic.go:334] "Generic (PLEG): container finished" podID="c23056d6-d516-4856-9309-e0523fd252d4" containerID="881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c" exitCode=0 Dec 08 20:22:43 crc kubenswrapper[4706]: I1208 20:22:43.426139 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htq42" event={"ID":"c23056d6-d516-4856-9309-e0523fd252d4","Type":"ContainerDied","Data":"881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c"} Dec 08 20:22:43 crc kubenswrapper[4706]: I1208 20:22:43.427623 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htq42" event={"ID":"c23056d6-d516-4856-9309-e0523fd252d4","Type":"ContainerStarted","Data":"c7df733b1d54f32f142a5de533febde4af2d3d11170dac69d1563cbfa138a6e1"} Dec 08 20:22:44 crc kubenswrapper[4706]: I1208 20:22:44.442958 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htq42" event={"ID":"c23056d6-d516-4856-9309-e0523fd252d4","Type":"ContainerStarted","Data":"0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8"} Dec 08 20:22:45 crc kubenswrapper[4706]: E1208 20:22:45.246555 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc23056d6_d516_4856_9309_e0523fd252d4.slice/crio-0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc23056d6_d516_4856_9309_e0523fd252d4.slice/crio-conmon-0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8.scope\": RecentStats: unable to find data in memory cache]" Dec 08 20:22:45 crc kubenswrapper[4706]: I1208 20:22:45.457720 4706 generic.go:334] "Generic (PLEG): container finished" podID="c23056d6-d516-4856-9309-e0523fd252d4" containerID="0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8" exitCode=0 Dec 08 20:22:45 crc kubenswrapper[4706]: I1208 20:22:45.457842 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htq42" event={"ID":"c23056d6-d516-4856-9309-e0523fd252d4","Type":"ContainerDied","Data":"0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8"} Dec 08 20:22:46 crc kubenswrapper[4706]: I1208 20:22:46.491345 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htq42" event={"ID":"c23056d6-d516-4856-9309-e0523fd252d4","Type":"ContainerStarted","Data":"96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb"} Dec 08 20:22:46 crc kubenswrapper[4706]: I1208 20:22:46.521489 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-htq42" podStartSLOduration=3.066070807 podStartE2EDuration="5.521464955s" podCreationTimestamp="2025-12-08 20:22:41 +0000 UTC" firstStartedPulling="2025-12-08 20:22:43.429017169 +0000 UTC m=+3646.071218172" lastFinishedPulling="2025-12-08 20:22:45.884411317 +0000 UTC m=+3648.526612320" observedRunningTime="2025-12-08 20:22:46.51035753 +0000 UTC m=+3649.152558543" watchObservedRunningTime="2025-12-08 20:22:46.521464955 +0000 UTC m=+3649.163665958" Dec 08 20:22:52 crc kubenswrapper[4706]: I1208 20:22:52.013337 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:52 crc kubenswrapper[4706]: I1208 20:22:52.014501 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:52 crc kubenswrapper[4706]: I1208 20:22:52.077127 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:52 crc kubenswrapper[4706]: I1208 20:22:52.621866 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:52 crc kubenswrapper[4706]: I1208 20:22:52.690986 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-htq42"] Dec 08 20:22:54 crc kubenswrapper[4706]: I1208 20:22:54.573147 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-htq42" podUID="c23056d6-d516-4856-9309-e0523fd252d4" containerName="registry-server" containerID="cri-o://96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb" gracePeriod=2 Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.386039 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.451237 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-catalog-content\") pod \"c23056d6-d516-4856-9309-e0523fd252d4\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.451531 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rlhz\" (UniqueName: \"kubernetes.io/projected/c23056d6-d516-4856-9309-e0523fd252d4-kube-api-access-4rlhz\") pod \"c23056d6-d516-4856-9309-e0523fd252d4\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.451750 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-utilities\") pod \"c23056d6-d516-4856-9309-e0523fd252d4\" (UID: \"c23056d6-d516-4856-9309-e0523fd252d4\") " Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.455236 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-utilities" (OuterVolumeSpecName: "utilities") pod "c23056d6-d516-4856-9309-e0523fd252d4" (UID: "c23056d6-d516-4856-9309-e0523fd252d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.460863 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c23056d6-d516-4856-9309-e0523fd252d4-kube-api-access-4rlhz" (OuterVolumeSpecName: "kube-api-access-4rlhz") pod "c23056d6-d516-4856-9309-e0523fd252d4" (UID: "c23056d6-d516-4856-9309-e0523fd252d4"). InnerVolumeSpecName "kube-api-access-4rlhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.532098 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c23056d6-d516-4856-9309-e0523fd252d4" (UID: "c23056d6-d516-4856-9309-e0523fd252d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.554402 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.554461 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23056d6-d516-4856-9309-e0523fd252d4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.554481 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rlhz\" (UniqueName: \"kubernetes.io/projected/c23056d6-d516-4856-9309-e0523fd252d4-kube-api-access-4rlhz\") on node \"crc\" DevicePath \"\"" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.590210 4706 generic.go:334] "Generic (PLEG): container finished" podID="c23056d6-d516-4856-9309-e0523fd252d4" containerID="96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb" exitCode=0 Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.590287 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htq42" event={"ID":"c23056d6-d516-4856-9309-e0523fd252d4","Type":"ContainerDied","Data":"96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb"} Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.590325 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htq42" event={"ID":"c23056d6-d516-4856-9309-e0523fd252d4","Type":"ContainerDied","Data":"c7df733b1d54f32f142a5de533febde4af2d3d11170dac69d1563cbfa138a6e1"} Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.590349 4706 scope.go:117] "RemoveContainer" containerID="96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.591613 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-htq42" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.617692 4706 scope.go:117] "RemoveContainer" containerID="0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.662769 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-htq42"] Dec 08 20:22:55 crc kubenswrapper[4706]: E1208 20:22:55.678666 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc23056d6_d516_4856_9309_e0523fd252d4.slice\": RecentStats: unable to find data in memory cache]" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.679152 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-htq42"] Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.684616 4706 scope.go:117] "RemoveContainer" containerID="881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.749737 4706 scope.go:117] "RemoveContainer" containerID="96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb" Dec 08 20:22:55 crc kubenswrapper[4706]: E1208 20:22:55.751940 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb\": container with ID starting with 96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb not found: ID does not exist" containerID="96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.752019 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb"} err="failed to get container status \"96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb\": rpc error: code = NotFound desc = could not find container \"96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb\": container with ID starting with 96889fb8538b46a98dd872a70d95b7f587b964c58da112bba3025aba69a799eb not found: ID does not exist" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.752061 4706 scope.go:117] "RemoveContainer" containerID="0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8" Dec 08 20:22:55 crc kubenswrapper[4706]: E1208 20:22:55.752730 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8\": container with ID starting with 0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8 not found: ID does not exist" containerID="0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.752757 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8"} err="failed to get container status \"0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8\": rpc error: code = NotFound desc = could not find container \"0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8\": container with ID starting with 0660799b61d7f5cd1a1db7b1e0a8ebb2d90090b37848dd6335da1c4c6b0e8bb8 not found: ID does not exist" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.752779 4706 scope.go:117] "RemoveContainer" containerID="881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c" Dec 08 20:22:55 crc kubenswrapper[4706]: E1208 20:22:55.753440 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c\": container with ID starting with 881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c not found: ID does not exist" containerID="881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c" Dec 08 20:22:55 crc kubenswrapper[4706]: I1208 20:22:55.753465 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c"} err="failed to get container status \"881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c\": rpc error: code = NotFound desc = could not find container \"881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c\": container with ID starting with 881ce9f4d5236cda76a3fce2a8c9d4f72fc1a5397305357c506b0ffcb221b85c not found: ID does not exist" Dec 08 20:22:57 crc kubenswrapper[4706]: I1208 20:22:57.626773 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c23056d6-d516-4856-9309-e0523fd252d4" path="/var/lib/kubelet/pods/c23056d6-d516-4856-9309-e0523fd252d4/volumes" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.453661 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kwnfn"] Dec 08 20:23:08 crc kubenswrapper[4706]: E1208 20:23:08.455196 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23056d6-d516-4856-9309-e0523fd252d4" containerName="extract-utilities" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.455215 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23056d6-d516-4856-9309-e0523fd252d4" containerName="extract-utilities" Dec 08 20:23:08 crc kubenswrapper[4706]: E1208 20:23:08.455243 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23056d6-d516-4856-9309-e0523fd252d4" containerName="registry-server" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.455252 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23056d6-d516-4856-9309-e0523fd252d4" containerName="registry-server" Dec 08 20:23:08 crc kubenswrapper[4706]: E1208 20:23:08.455300 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23056d6-d516-4856-9309-e0523fd252d4" containerName="extract-content" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.455308 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23056d6-d516-4856-9309-e0523fd252d4" containerName="extract-content" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.455582 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c23056d6-d516-4856-9309-e0523fd252d4" containerName="registry-server" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.457745 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.472408 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kwnfn"] Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.620021 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-catalog-content\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.620233 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pjpv\" (UniqueName: \"kubernetes.io/projected/6ec377f2-5e17-46ba-aedb-383182dd9c05-kube-api-access-6pjpv\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.620436 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-utilities\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.723933 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-utilities\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.724024 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-catalog-content\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.724219 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pjpv\" (UniqueName: \"kubernetes.io/projected/6ec377f2-5e17-46ba-aedb-383182dd9c05-kube-api-access-6pjpv\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.724755 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-utilities\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.724864 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-catalog-content\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.756924 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pjpv\" (UniqueName: \"kubernetes.io/projected/6ec377f2-5e17-46ba-aedb-383182dd9c05-kube-api-access-6pjpv\") pod \"redhat-operators-kwnfn\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:08 crc kubenswrapper[4706]: I1208 20:23:08.809755 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:09 crc kubenswrapper[4706]: I1208 20:23:09.386603 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kwnfn"] Dec 08 20:23:10 crc kubenswrapper[4706]: I1208 20:23:10.419405 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwnfn" event={"ID":"6ec377f2-5e17-46ba-aedb-383182dd9c05","Type":"ContainerStarted","Data":"08d19812baa1eb987c677b898192510557f1dae544f2784453b39d7c3cfd45b8"} Dec 08 20:23:11 crc kubenswrapper[4706]: I1208 20:23:11.433921 4706 generic.go:334] "Generic (PLEG): container finished" podID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerID="01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912" exitCode=0 Dec 08 20:23:11 crc kubenswrapper[4706]: I1208 20:23:11.434139 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwnfn" event={"ID":"6ec377f2-5e17-46ba-aedb-383182dd9c05","Type":"ContainerDied","Data":"01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912"} Dec 08 20:23:12 crc kubenswrapper[4706]: I1208 20:23:12.450748 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwnfn" event={"ID":"6ec377f2-5e17-46ba-aedb-383182dd9c05","Type":"ContainerStarted","Data":"238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00"} Dec 08 20:23:12 crc kubenswrapper[4706]: I1208 20:23:12.788313 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-qtnlw_4c42f8b9-a85c-4a45-a8f8-4a43e561a14e/kube-rbac-proxy/0.log" Dec 08 20:23:12 crc kubenswrapper[4706]: I1208 20:23:12.865953 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-qtnlw_4c42f8b9-a85c-4a45-a8f8-4a43e561a14e/manager/0.log" Dec 08 20:23:13 crc kubenswrapper[4706]: I1208 20:23:13.169154 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-j2mjl_e5899e67-6618-43f5-8454-fddfae494e4a/kube-rbac-proxy/0.log" Dec 08 20:23:13 crc kubenswrapper[4706]: I1208 20:23:13.362866 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-j2mjl_e5899e67-6618-43f5-8454-fddfae494e4a/manager/0.log" Dec 08 20:23:13 crc kubenswrapper[4706]: I1208 20:23:13.726039 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/util/0.log" Dec 08 20:23:13 crc kubenswrapper[4706]: I1208 20:23:13.873349 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/pull/0.log" Dec 08 20:23:13 crc kubenswrapper[4706]: I1208 20:23:13.939221 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/util/0.log" Dec 08 20:23:13 crc kubenswrapper[4706]: I1208 20:23:13.940314 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/pull/0.log" Dec 08 20:23:14 crc kubenswrapper[4706]: I1208 20:23:14.338194 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/pull/0.log" Dec 08 20:23:14 crc kubenswrapper[4706]: I1208 20:23:14.389987 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/extract/0.log" Dec 08 20:23:14 crc kubenswrapper[4706]: I1208 20:23:14.399339 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/util/0.log" Dec 08 20:23:14 crc kubenswrapper[4706]: I1208 20:23:14.611868 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-7fvjs_0f144ce9-7cc0-41f3-84ea-654579d32c0b/kube-rbac-proxy/0.log" Dec 08 20:23:14 crc kubenswrapper[4706]: I1208 20:23:14.677445 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-7fvjs_0f144ce9-7cc0-41f3-84ea-654579d32c0b/manager/0.log" Dec 08 20:23:14 crc kubenswrapper[4706]: I1208 20:23:14.794107 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-b7m9l_b02b307a-7d4c-4d20-b388-18f6fc6ec0ee/kube-rbac-proxy/0.log" Dec 08 20:23:14 crc kubenswrapper[4706]: I1208 20:23:14.982463 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-b7m9l_b02b307a-7d4c-4d20-b388-18f6fc6ec0ee/manager/0.log" Dec 08 20:23:15 crc kubenswrapper[4706]: I1208 20:23:15.129869 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-cmg9l_2acd6f54-87a9-4e50-859c-42a7c8e9d098/kube-rbac-proxy/0.log" Dec 08 20:23:15 crc kubenswrapper[4706]: I1208 20:23:15.167548 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-cmg9l_2acd6f54-87a9-4e50-859c-42a7c8e9d098/manager/0.log" Dec 08 20:23:15 crc kubenswrapper[4706]: I1208 20:23:15.711449 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cggng_1290df78-3e2c-494e-a264-c495d80c0bef/kube-rbac-proxy/0.log" Dec 08 20:23:15 crc kubenswrapper[4706]: I1208 20:23:15.734502 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cggng_1290df78-3e2c-494e-a264-c495d80c0bef/manager/0.log" Dec 08 20:23:16 crc kubenswrapper[4706]: I1208 20:23:16.126985 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-qmcdh_ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3/kube-rbac-proxy/0.log" Dec 08 20:23:16 crc kubenswrapper[4706]: I1208 20:23:16.353346 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5cwxk_42d526f1-2349-4aef-80fa-a6e59f3630e9/kube-rbac-proxy/0.log" Dec 08 20:23:16 crc kubenswrapper[4706]: I1208 20:23:16.412349 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5cwxk_42d526f1-2349-4aef-80fa-a6e59f3630e9/manager/0.log" Dec 08 20:23:16 crc kubenswrapper[4706]: I1208 20:23:16.505539 4706 generic.go:334] "Generic (PLEG): container finished" podID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerID="238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00" exitCode=0 Dec 08 20:23:16 crc kubenswrapper[4706]: I1208 20:23:16.505611 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwnfn" event={"ID":"6ec377f2-5e17-46ba-aedb-383182dd9c05","Type":"ContainerDied","Data":"238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00"} Dec 08 20:23:16 crc kubenswrapper[4706]: I1208 20:23:16.614982 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-qmcdh_ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3/manager/0.log" Dec 08 20:23:16 crc kubenswrapper[4706]: I1208 20:23:16.692612 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kg2wq_fab26e20-8c1c-4a9e-b7c4-8d4815a1568e/kube-rbac-proxy/0.log" Dec 08 20:23:16 crc kubenswrapper[4706]: I1208 20:23:16.772627 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kg2wq_fab26e20-8c1c-4a9e-b7c4-8d4815a1568e/manager/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.016552 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-pfq9b_53d61380-3a1e-45e5-a039-9d339bcdb9ea/manager/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.190956 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-pfq9b_53d61380-3a1e-45e5-a039-9d339bcdb9ea/kube-rbac-proxy/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.408355 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-l8qxc_8ddb889c-1432-4225-a110-e0aabc0ce3bc/kube-rbac-proxy/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.478784 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-5xgcd_1f3c9d68-2093-45a5-886e-f19c43b3e8e9/kube-rbac-proxy/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.512587 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-l8qxc_8ddb889c-1432-4225-a110-e0aabc0ce3bc/manager/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.532071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwnfn" event={"ID":"6ec377f2-5e17-46ba-aedb-383182dd9c05","Type":"ContainerStarted","Data":"2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7"} Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.553626 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-5xgcd_1f3c9d68-2093-45a5-886e-f19c43b3e8e9/manager/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.698481 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-58rkn_6da8aab3-b971-413d-8557-cdfad2c95d2c/kube-rbac-proxy/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.874320 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5ndn8_eedc4c77-93bc-40b9-ba3c-e7222d8828b5/kube-rbac-proxy/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.900651 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-58rkn_6da8aab3-b971-413d-8557-cdfad2c95d2c/manager/0.log" Dec 08 20:23:17 crc kubenswrapper[4706]: I1208 20:23:17.982283 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5ndn8_eedc4c77-93bc-40b9-ba3c-e7222d8828b5/manager/0.log" Dec 08 20:23:18 crc kubenswrapper[4706]: I1208 20:23:18.155547 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f6nqtx_4fca7ad0-1e55-4b7a-bee5-08521a1836a2/manager/0.log" Dec 08 20:23:18 crc kubenswrapper[4706]: I1208 20:23:18.181538 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f6nqtx_4fca7ad0-1e55-4b7a-bee5-08521a1836a2/kube-rbac-proxy/0.log" Dec 08 20:23:18 crc kubenswrapper[4706]: I1208 20:23:18.732755 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d96469866-85q7p_ea53cbe8-559b-45be-8ae8-25aeac554de4/operator/0.log" Dec 08 20:23:18 crc kubenswrapper[4706]: I1208 20:23:18.809955 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:18 crc kubenswrapper[4706]: I1208 20:23:18.810015 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:19 crc kubenswrapper[4706]: I1208 20:23:19.074041 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-c5fr6_ca6c738d-aa2e-4be3-9827-fc03af84458d/registry-server/0.log" Dec 08 20:23:19 crc kubenswrapper[4706]: I1208 20:23:19.461763 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-vpwqp_ec6df814-0549-46a9-a9f3-2f94083bda52/kube-rbac-proxy/0.log" Dec 08 20:23:19 crc kubenswrapper[4706]: I1208 20:23:19.491478 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-b89665ccc-lvggm_61674e93-1252-4e28-ac66-b2a701745112/manager/0.log" Dec 08 20:23:19 crc kubenswrapper[4706]: I1208 20:23:19.530244 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-vpwqp_ec6df814-0549-46a9-a9f3-2f94083bda52/manager/0.log" Dec 08 20:23:19 crc kubenswrapper[4706]: I1208 20:23:19.536601 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nbcmr_22335991-2ba5-4fba-a1d1-b391e4767961/kube-rbac-proxy/0.log" Dec 08 20:23:19 crc kubenswrapper[4706]: I1208 20:23:19.893600 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kwnfn" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="registry-server" probeResult="failure" output=< Dec 08 20:23:19 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 20:23:19 crc kubenswrapper[4706]: > Dec 08 20:23:19 crc kubenswrapper[4706]: I1208 20:23:19.899621 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nbcmr_22335991-2ba5-4fba-a1d1-b391e4767961/manager/0.log" Dec 08 20:23:19 crc kubenswrapper[4706]: I1208 20:23:19.991865 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-7htmr_e9c55dcb-b67a-43f9-9a80-8c38e7f94034/kube-rbac-proxy/0.log" Dec 08 20:23:20 crc kubenswrapper[4706]: I1208 20:23:20.000458 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-bqsg2_9cca1f0f-9037-49d5-b8da-392421a67702/operator/0.log" Dec 08 20:23:20 crc kubenswrapper[4706]: I1208 20:23:20.188594 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-7htmr_e9c55dcb-b67a-43f9-9a80-8c38e7f94034/manager/0.log" Dec 08 20:23:20 crc kubenswrapper[4706]: I1208 20:23:20.272127 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5cbd74b459-5l5k7_beda813f-668e-45e8-8d1c-7c72caa6f294/kube-rbac-proxy/0.log" Dec 08 20:23:20 crc kubenswrapper[4706]: I1208 20:23:20.464092 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-qfsq9_375c721a-f040-42a8-9b43-f609258063a1/kube-rbac-proxy/0.log" Dec 08 20:23:20 crc kubenswrapper[4706]: I1208 20:23:20.596798 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5cbd74b459-5l5k7_beda813f-668e-45e8-8d1c-7c72caa6f294/manager/0.log" Dec 08 20:23:20 crc kubenswrapper[4706]: I1208 20:23:20.625131 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-qfsq9_375c721a-f040-42a8-9b43-f609258063a1/manager/0.log" Dec 08 20:23:20 crc kubenswrapper[4706]: I1208 20:23:20.720828 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-g6gls_ed8cb084-a66c-4fc0-afa1-068e58c80680/kube-rbac-proxy/0.log" Dec 08 20:23:20 crc kubenswrapper[4706]: I1208 20:23:20.804207 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-g6gls_ed8cb084-a66c-4fc0-afa1-068e58c80680/manager/0.log" Dec 08 20:23:28 crc kubenswrapper[4706]: I1208 20:23:28.885166 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:28 crc kubenswrapper[4706]: I1208 20:23:28.923127 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kwnfn" podStartSLOduration=15.394243069 podStartE2EDuration="20.923095491s" podCreationTimestamp="2025-12-08 20:23:08 +0000 UTC" firstStartedPulling="2025-12-08 20:23:11.438147119 +0000 UTC m=+3674.080348122" lastFinishedPulling="2025-12-08 20:23:16.966999541 +0000 UTC m=+3679.609200544" observedRunningTime="2025-12-08 20:23:17.55861422 +0000 UTC m=+3680.200815253" watchObservedRunningTime="2025-12-08 20:23:28.923095491 +0000 UTC m=+3691.565296494" Dec 08 20:23:28 crc kubenswrapper[4706]: I1208 20:23:28.973163 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:29 crc kubenswrapper[4706]: I1208 20:23:29.138380 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kwnfn"] Dec 08 20:23:30 crc kubenswrapper[4706]: I1208 20:23:30.696720 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kwnfn" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="registry-server" containerID="cri-o://2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7" gracePeriod=2 Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.462790 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.517573 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-utilities\") pod \"6ec377f2-5e17-46ba-aedb-383182dd9c05\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.517688 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-catalog-content\") pod \"6ec377f2-5e17-46ba-aedb-383182dd9c05\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.517754 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pjpv\" (UniqueName: \"kubernetes.io/projected/6ec377f2-5e17-46ba-aedb-383182dd9c05-kube-api-access-6pjpv\") pod \"6ec377f2-5e17-46ba-aedb-383182dd9c05\" (UID: \"6ec377f2-5e17-46ba-aedb-383182dd9c05\") " Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.518797 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-utilities" (OuterVolumeSpecName: "utilities") pod "6ec377f2-5e17-46ba-aedb-383182dd9c05" (UID: "6ec377f2-5e17-46ba-aedb-383182dd9c05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.534619 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ec377f2-5e17-46ba-aedb-383182dd9c05-kube-api-access-6pjpv" (OuterVolumeSpecName: "kube-api-access-6pjpv") pod "6ec377f2-5e17-46ba-aedb-383182dd9c05" (UID: "6ec377f2-5e17-46ba-aedb-383182dd9c05"). InnerVolumeSpecName "kube-api-access-6pjpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.627928 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.627970 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pjpv\" (UniqueName: \"kubernetes.io/projected/6ec377f2-5e17-46ba-aedb-383182dd9c05-kube-api-access-6pjpv\") on node \"crc\" DevicePath \"\"" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.719848 4706 generic.go:334] "Generic (PLEG): container finished" podID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerID="2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7" exitCode=0 Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.719909 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwnfn" event={"ID":"6ec377f2-5e17-46ba-aedb-383182dd9c05","Type":"ContainerDied","Data":"2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7"} Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.719952 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwnfn" event={"ID":"6ec377f2-5e17-46ba-aedb-383182dd9c05","Type":"ContainerDied","Data":"08d19812baa1eb987c677b898192510557f1dae544f2784453b39d7c3cfd45b8"} Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.719975 4706 scope.go:117] "RemoveContainer" containerID="2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.720149 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kwnfn" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.745547 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ec377f2-5e17-46ba-aedb-383182dd9c05" (UID: "6ec377f2-5e17-46ba-aedb-383182dd9c05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.751049 4706 scope.go:117] "RemoveContainer" containerID="238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.786169 4706 scope.go:117] "RemoveContainer" containerID="01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.835829 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec377f2-5e17-46ba-aedb-383182dd9c05-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.854234 4706 scope.go:117] "RemoveContainer" containerID="2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7" Dec 08 20:23:31 crc kubenswrapper[4706]: E1208 20:23:31.856231 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7\": container with ID starting with 2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7 not found: ID does not exist" containerID="2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.856285 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7"} err="failed to get container status \"2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7\": rpc error: code = NotFound desc = could not find container \"2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7\": container with ID starting with 2cbb1c07749744f761bda80bac14b44e43923a9d09867ce914ea8e035c3229f7 not found: ID does not exist" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.856311 4706 scope.go:117] "RemoveContainer" containerID="238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00" Dec 08 20:23:31 crc kubenswrapper[4706]: E1208 20:23:31.857109 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00\": container with ID starting with 238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00 not found: ID does not exist" containerID="238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.857272 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00"} err="failed to get container status \"238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00\": rpc error: code = NotFound desc = could not find container \"238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00\": container with ID starting with 238b24ccf3e4b58d6f8de92e903719c0792cbf1a33f1226369f2d039065e7b00 not found: ID does not exist" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.857413 4706 scope.go:117] "RemoveContainer" containerID="01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912" Dec 08 20:23:31 crc kubenswrapper[4706]: E1208 20:23:31.858619 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912\": container with ID starting with 01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912 not found: ID does not exist" containerID="01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912" Dec 08 20:23:31 crc kubenswrapper[4706]: I1208 20:23:31.858694 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912"} err="failed to get container status \"01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912\": rpc error: code = NotFound desc = could not find container \"01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912\": container with ID starting with 01c0bfd4381d0d1e9b2beebdf0295e8f25deba2723b6d4c74c081cb32ad88912 not found: ID does not exist" Dec 08 20:23:32 crc kubenswrapper[4706]: I1208 20:23:32.064561 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kwnfn"] Dec 08 20:23:32 crc kubenswrapper[4706]: I1208 20:23:32.088686 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kwnfn"] Dec 08 20:23:33 crc kubenswrapper[4706]: I1208 20:23:33.647147 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" path="/var/lib/kubelet/pods/6ec377f2-5e17-46ba-aedb-383182dd9c05/volumes" Dec 08 20:23:49 crc kubenswrapper[4706]: I1208 20:23:49.782679 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mxb2z_878ce0cf-8408-4bd3-90cc-948a66892c07/control-plane-machine-set-operator/0.log" Dec 08 20:23:50 crc kubenswrapper[4706]: I1208 20:23:50.053059 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j6s78_48960e60-2160-4afa-98ca-e67603fc9545/machine-api-operator/0.log" Dec 08 20:23:50 crc kubenswrapper[4706]: I1208 20:23:50.074492 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j6s78_48960e60-2160-4afa-98ca-e67603fc9545/kube-rbac-proxy/0.log" Dec 08 20:24:07 crc kubenswrapper[4706]: I1208 20:24:07.241953 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-c6s5g_8bf48929-3e5e-4fc7-a559-0ddcc2707cda/cert-manager-controller/0.log" Dec 08 20:24:07 crc kubenswrapper[4706]: I1208 20:24:07.424974 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-hzrrx_aefedb16-e712-45a1-8c7b-cd34002fd0bb/cert-manager-cainjector/0.log" Dec 08 20:24:07 crc kubenswrapper[4706]: I1208 20:24:07.528073 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-588tt_c1b92b7d-f282-496b-b69f-f55c0b9baec2/cert-manager-webhook/0.log" Dec 08 20:24:25 crc kubenswrapper[4706]: I1208 20:24:25.374044 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-224bh_0b6da606-8fe3-4228-81f7-533d8155c6f0/nmstate-console-plugin/0.log" Dec 08 20:24:25 crc kubenswrapper[4706]: I1208 20:24:25.809783 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-llfnx_05ea3b8c-286e-48e7-b232-ac550fcf165e/kube-rbac-proxy/0.log" Dec 08 20:24:25 crc kubenswrapper[4706]: I1208 20:24:25.819819 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-zqns4_1287503f-533b-4f53-b9b2-6792ad891790/nmstate-handler/0.log" Dec 08 20:24:25 crc kubenswrapper[4706]: I1208 20:24:25.981044 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-llfnx_05ea3b8c-286e-48e7-b232-ac550fcf165e/nmstate-metrics/0.log" Dec 08 20:24:26 crc kubenswrapper[4706]: I1208 20:24:26.151928 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-9jz6k_f56b4031-2db2-4d47-9299-9829b3963664/nmstate-operator/0.log" Dec 08 20:24:26 crc kubenswrapper[4706]: I1208 20:24:26.238313 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-kwnzx_aea9659b-8952-4f71-a8d5-970bfd9fffe0/nmstate-webhook/0.log" Dec 08 20:24:44 crc kubenswrapper[4706]: I1208 20:24:44.661061 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55474f77b6-v6n6q_b600c67d-b283-41ed-85f1-d19f3671f40f/kube-rbac-proxy/0.log" Dec 08 20:24:44 crc kubenswrapper[4706]: I1208 20:24:44.678747 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55474f77b6-v6n6q_b600c67d-b283-41ed-85f1-d19f3671f40f/manager/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.127035 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-9k84w_00d01f1b-f2d8-467d-94a3-e9a73966d334/kube-rbac-proxy/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.209683 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-9k84w_00d01f1b-f2d8-467d-94a3-e9a73966d334/controller/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.343030 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-frr-files/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.563012 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-frr-files/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.600399 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-reloader/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.601252 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-metrics/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.630488 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-reloader/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.817900 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-reloader/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.951731 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-metrics/0.log" Dec 08 20:25:02 crc kubenswrapper[4706]: I1208 20:25:02.990405 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-metrics/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.173501 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-frr-files/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.354538 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-reloader/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.395729 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-frr-files/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.444950 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-metrics/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.486425 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/controller/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.722158 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/frr-metrics/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.783886 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/kube-rbac-proxy/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.848115 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/kube-rbac-proxy-frr/0.log" Dec 08 20:25:03 crc kubenswrapper[4706]: I1208 20:25:03.944848 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/reloader/0.log" Dec 08 20:25:04 crc kubenswrapper[4706]: I1208 20:25:04.085805 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-mdfmz_257f3480-2853-4319-b4c2-7eb65e6bb097/frr-k8s-webhook-server/0.log" Dec 08 20:25:04 crc kubenswrapper[4706]: I1208 20:25:04.353147 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-65c6ff64d8-bdjmj_fdc00423-9550-4ff3-84d9-78eeb50e29ea/manager/0.log" Dec 08 20:25:04 crc kubenswrapper[4706]: I1208 20:25:04.538668 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-75457555df-ll7rx_4ef64783-e4c8-4c02-884f-3dba766387b8/webhook-server/0.log" Dec 08 20:25:04 crc kubenswrapper[4706]: I1208 20:25:04.744092 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-49kgz_1398dd3e-9aac-4f48-b927-1862ed71c544/kube-rbac-proxy/0.log" Dec 08 20:25:05 crc kubenswrapper[4706]: I1208 20:25:05.301364 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/frr/0.log" Dec 08 20:25:05 crc kubenswrapper[4706]: I1208 20:25:05.368710 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-49kgz_1398dd3e-9aac-4f48-b927-1862ed71c544/speaker/0.log" Dec 08 20:25:05 crc kubenswrapper[4706]: I1208 20:25:05.836588 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:25:05 crc kubenswrapper[4706]: I1208 20:25:05.837106 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:25:20 crc kubenswrapper[4706]: I1208 20:25:20.062870 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/util/0.log" Dec 08 20:25:20 crc kubenswrapper[4706]: I1208 20:25:20.266993 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/util/0.log" Dec 08 20:25:20 crc kubenswrapper[4706]: I1208 20:25:20.322892 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/pull/0.log" Dec 08 20:25:20 crc kubenswrapper[4706]: I1208 20:25:20.323117 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/pull/0.log" Dec 08 20:25:20 crc kubenswrapper[4706]: I1208 20:25:20.511231 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/pull/0.log" Dec 08 20:25:20 crc kubenswrapper[4706]: I1208 20:25:20.515550 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/util/0.log" Dec 08 20:25:20 crc kubenswrapper[4706]: I1208 20:25:20.542365 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/extract/0.log" Dec 08 20:25:20 crc kubenswrapper[4706]: I1208 20:25:20.736929 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/util/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.035179 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/util/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.037523 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/pull/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.077876 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/pull/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.287708 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/pull/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.334100 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/util/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.401141 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/extract/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.549363 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/util/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.733631 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/util/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.758651 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/pull/0.log" Dec 08 20:25:21 crc kubenswrapper[4706]: I1208 20:25:21.769599 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/pull/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.019776 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/extract/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.065178 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/util/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.099498 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/pull/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.226130 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/util/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.424764 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/util/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.434542 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/pull/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.454104 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/pull/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.648204 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/util/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.692218 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/extract/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.693330 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/pull/0.log" Dec 08 20:25:22 crc kubenswrapper[4706]: I1208 20:25:22.879848 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-utilities/0.log" Dec 08 20:25:23 crc kubenswrapper[4706]: I1208 20:25:23.041709 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-content/0.log" Dec 08 20:25:23 crc kubenswrapper[4706]: I1208 20:25:23.080311 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-utilities/0.log" Dec 08 20:25:23 crc kubenswrapper[4706]: I1208 20:25:23.139418 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-content/0.log" Dec 08 20:25:23 crc kubenswrapper[4706]: I1208 20:25:23.332066 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-content/0.log" Dec 08 20:25:23 crc kubenswrapper[4706]: I1208 20:25:23.418242 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-utilities/0.log" Dec 08 20:25:23 crc kubenswrapper[4706]: I1208 20:25:23.633943 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-utilities/0.log" Dec 08 20:25:23 crc kubenswrapper[4706]: I1208 20:25:23.936945 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-utilities/0.log" Dec 08 20:25:23 crc kubenswrapper[4706]: I1208 20:25:23.998706 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-content/0.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.029961 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-content/0.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.031446 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/registry-server/0.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.244470 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-utilities/0.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.258469 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-content/0.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.388332 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/3.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.540163 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/2.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.718571 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-utilities/0.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.927159 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-content/0.log" Dec 08 20:25:24 crc kubenswrapper[4706]: I1208 20:25:24.932248 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-content/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.022960 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-utilities/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.109073 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/registry-server/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.259798 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-utilities/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.309294 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-content/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.397797 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-utilities/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.431928 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/registry-server/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.690471 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-content/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.690556 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-utilities/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.700632 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-content/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.869081 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-content/0.log" Dec 08 20:25:25 crc kubenswrapper[4706]: I1208 20:25:25.869461 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-utilities/0.log" Dec 08 20:25:26 crc kubenswrapper[4706]: I1208 20:25:26.399969 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/registry-server/0.log" Dec 08 20:25:35 crc kubenswrapper[4706]: I1208 20:25:35.835895 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:25:35 crc kubenswrapper[4706]: I1208 20:25:35.836909 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:25:40 crc kubenswrapper[4706]: I1208 20:25:40.695090 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-fntdr_90a15b07-3ae8-4265-a92b-d9dd052211e2/prometheus-operator/0.log" Dec 08 20:25:40 crc kubenswrapper[4706]: I1208 20:25:40.775187 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_3f251b1b-6576-4421-aa76-04a26d0e1302/prometheus-operator-admission-webhook/0.log" Dec 08 20:25:40 crc kubenswrapper[4706]: I1208 20:25:40.876457 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c/prometheus-operator-admission-webhook/0.log" Dec 08 20:25:41 crc kubenswrapper[4706]: I1208 20:25:41.080293 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-4xlqn_3cbb9b0b-a2c7-47f8-92d2-de4829687a45/operator/0.log" Dec 08 20:25:41 crc kubenswrapper[4706]: I1208 20:25:41.190920 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-fqdzc_a8e2aac6-1870-4ddc-8650-c31ded55337f/perses-operator/0.log" Dec 08 20:25:55 crc kubenswrapper[4706]: I1208 20:25:55.081441 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55474f77b6-v6n6q_b600c67d-b283-41ed-85f1-d19f3671f40f/kube-rbac-proxy/0.log" Dec 08 20:25:55 crc kubenswrapper[4706]: I1208 20:25:55.166794 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55474f77b6-v6n6q_b600c67d-b283-41ed-85f1-d19f3671f40f/manager/0.log" Dec 08 20:26:05 crc kubenswrapper[4706]: I1208 20:26:05.836175 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:26:05 crc kubenswrapper[4706]: I1208 20:26:05.837100 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:26:05 crc kubenswrapper[4706]: I1208 20:26:05.837159 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 20:26:05 crc kubenswrapper[4706]: I1208 20:26:05.838204 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 20:26:05 crc kubenswrapper[4706]: I1208 20:26:05.838253 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" gracePeriod=600 Dec 08 20:26:05 crc kubenswrapper[4706]: E1208 20:26:05.963846 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:26:06 crc kubenswrapper[4706]: I1208 20:26:06.078415 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" exitCode=0 Dec 08 20:26:06 crc kubenswrapper[4706]: I1208 20:26:06.078479 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451"} Dec 08 20:26:06 crc kubenswrapper[4706]: I1208 20:26:06.078531 4706 scope.go:117] "RemoveContainer" containerID="ae3b6b90e663cc0e30fdffcaf7a2453642e4c1963568b4eafa4406e240816b79" Dec 08 20:26:06 crc kubenswrapper[4706]: I1208 20:26:06.079740 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:26:06 crc kubenswrapper[4706]: E1208 20:26:06.080113 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:26:16 crc kubenswrapper[4706]: I1208 20:26:16.609577 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:26:16 crc kubenswrapper[4706]: E1208 20:26:16.610713 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:26:21 crc kubenswrapper[4706]: E1208 20:26:21.689917 4706 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:36144->38.102.83.113:43781: write tcp 38.102.83.113:36144->38.102.83.113:43781: write: broken pipe Dec 08 20:26:27 crc kubenswrapper[4706]: I1208 20:26:27.608780 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:26:27 crc kubenswrapper[4706]: E1208 20:26:27.610200 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:26:38 crc kubenswrapper[4706]: I1208 20:26:38.609705 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:26:38 crc kubenswrapper[4706]: E1208 20:26:38.610827 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:26:51 crc kubenswrapper[4706]: I1208 20:26:51.621520 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:26:51 crc kubenswrapper[4706]: E1208 20:26:51.624545 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:27:06 crc kubenswrapper[4706]: I1208 20:27:06.609624 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:27:06 crc kubenswrapper[4706]: E1208 20:27:06.611048 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:27:14 crc kubenswrapper[4706]: I1208 20:27:14.029615 4706 scope.go:117] "RemoveContainer" containerID="7c42cf55b53cc00cf6cff87d096e0cba843a30019dc28fcf6fdba8bee44f3718" Dec 08 20:27:20 crc kubenswrapper[4706]: I1208 20:27:20.610891 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:27:20 crc kubenswrapper[4706]: E1208 20:27:20.612329 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:27:35 crc kubenswrapper[4706]: I1208 20:27:35.609453 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:27:35 crc kubenswrapper[4706]: E1208 20:27:35.610613 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:27:46 crc kubenswrapper[4706]: I1208 20:27:46.608800 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:27:46 crc kubenswrapper[4706]: E1208 20:27:46.609801 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:28:00 crc kubenswrapper[4706]: I1208 20:28:00.609345 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:28:00 crc kubenswrapper[4706]: E1208 20:28:00.610638 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:28:01 crc kubenswrapper[4706]: I1208 20:28:01.742421 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b14077-f0ed-4648-af39-7517847cbae5" containerID="ba604c294f41059bfc2f36e4f782d45b0b839a7e0ca4b04f6446ed37491c226d" exitCode=0 Dec 08 20:28:01 crc kubenswrapper[4706]: I1208 20:28:01.743019 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" event={"ID":"d1b14077-f0ed-4648-af39-7517847cbae5","Type":"ContainerDied","Data":"ba604c294f41059bfc2f36e4f782d45b0b839a7e0ca4b04f6446ed37491c226d"} Dec 08 20:28:01 crc kubenswrapper[4706]: I1208 20:28:01.746519 4706 scope.go:117] "RemoveContainer" containerID="ba604c294f41059bfc2f36e4f782d45b0b839a7e0ca4b04f6446ed37491c226d" Dec 08 20:28:02 crc kubenswrapper[4706]: I1208 20:28:02.657650 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jfnm9_must-gather-r4pxc_d1b14077-f0ed-4648-af39-7517847cbae5/gather/0.log" Dec 08 20:28:10 crc kubenswrapper[4706]: I1208 20:28:10.619026 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jfnm9/must-gather-r4pxc"] Dec 08 20:28:10 crc kubenswrapper[4706]: I1208 20:28:10.620332 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" podUID="d1b14077-f0ed-4648-af39-7517847cbae5" containerName="copy" containerID="cri-o://aa15767d6300fec1431c5b3c9b7258d556cf2287209048f3ed9daa84eae1873f" gracePeriod=2 Dec 08 20:28:10 crc kubenswrapper[4706]: I1208 20:28:10.638237 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jfnm9/must-gather-r4pxc"] Dec 08 20:28:10 crc kubenswrapper[4706]: I1208 20:28:10.890109 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jfnm9_must-gather-r4pxc_d1b14077-f0ed-4648-af39-7517847cbae5/copy/0.log" Dec 08 20:28:10 crc kubenswrapper[4706]: I1208 20:28:10.891494 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b14077-f0ed-4648-af39-7517847cbae5" containerID="aa15767d6300fec1431c5b3c9b7258d556cf2287209048f3ed9daa84eae1873f" exitCode=143 Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.307355 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jfnm9_must-gather-r4pxc_d1b14077-f0ed-4648-af39-7517847cbae5/copy/0.log" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.308320 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.335582 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx6g8\" (UniqueName: \"kubernetes.io/projected/d1b14077-f0ed-4648-af39-7517847cbae5-kube-api-access-cx6g8\") pod \"d1b14077-f0ed-4648-af39-7517847cbae5\" (UID: \"d1b14077-f0ed-4648-af39-7517847cbae5\") " Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.335903 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d1b14077-f0ed-4648-af39-7517847cbae5-must-gather-output\") pod \"d1b14077-f0ed-4648-af39-7517847cbae5\" (UID: \"d1b14077-f0ed-4648-af39-7517847cbae5\") " Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.345054 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b14077-f0ed-4648-af39-7517847cbae5-kube-api-access-cx6g8" (OuterVolumeSpecName: "kube-api-access-cx6g8") pod "d1b14077-f0ed-4648-af39-7517847cbae5" (UID: "d1b14077-f0ed-4648-af39-7517847cbae5"). InnerVolumeSpecName "kube-api-access-cx6g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.440954 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx6g8\" (UniqueName: \"kubernetes.io/projected/d1b14077-f0ed-4648-af39-7517847cbae5-kube-api-access-cx6g8\") on node \"crc\" DevicePath \"\"" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.540866 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1b14077-f0ed-4648-af39-7517847cbae5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d1b14077-f0ed-4648-af39-7517847cbae5" (UID: "d1b14077-f0ed-4648-af39-7517847cbae5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.543073 4706 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d1b14077-f0ed-4648-af39-7517847cbae5-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.634589 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b14077-f0ed-4648-af39-7517847cbae5" path="/var/lib/kubelet/pods/d1b14077-f0ed-4648-af39-7517847cbae5/volumes" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.905363 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jfnm9_must-gather-r4pxc_d1b14077-f0ed-4648-af39-7517847cbae5/copy/0.log" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.906613 4706 scope.go:117] "RemoveContainer" containerID="aa15767d6300fec1431c5b3c9b7258d556cf2287209048f3ed9daa84eae1873f" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.906849 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfnm9/must-gather-r4pxc" Dec 08 20:28:11 crc kubenswrapper[4706]: I1208 20:28:11.941084 4706 scope.go:117] "RemoveContainer" containerID="ba604c294f41059bfc2f36e4f782d45b0b839a7e0ca4b04f6446ed37491c226d" Dec 08 20:28:13 crc kubenswrapper[4706]: I1208 20:28:13.611153 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:28:13 crc kubenswrapper[4706]: E1208 20:28:13.612082 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:28:14 crc kubenswrapper[4706]: I1208 20:28:14.154064 4706 scope.go:117] "RemoveContainer" containerID="f31f54de62a095892a4a75e8139f828219b3f2f55d50859ce7900c7b01a85bc3" Dec 08 20:28:24 crc kubenswrapper[4706]: I1208 20:28:24.608647 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:28:24 crc kubenswrapper[4706]: E1208 20:28:24.609883 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:28:37 crc kubenswrapper[4706]: I1208 20:28:37.618008 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:28:37 crc kubenswrapper[4706]: E1208 20:28:37.619220 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:28:51 crc kubenswrapper[4706]: I1208 20:28:51.608443 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:28:51 crc kubenswrapper[4706]: E1208 20:28:51.609571 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:29:02 crc kubenswrapper[4706]: I1208 20:29:02.609372 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:29:02 crc kubenswrapper[4706]: E1208 20:29:02.610784 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:29:17 crc kubenswrapper[4706]: I1208 20:29:17.635526 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:29:17 crc kubenswrapper[4706]: E1208 20:29:17.637682 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:29:31 crc kubenswrapper[4706]: I1208 20:29:31.609552 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:29:31 crc kubenswrapper[4706]: E1208 20:29:31.610791 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:29:42 crc kubenswrapper[4706]: I1208 20:29:42.607780 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:29:42 crc kubenswrapper[4706]: E1208 20:29:42.608450 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:29:53 crc kubenswrapper[4706]: I1208 20:29:53.609583 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:29:53 crc kubenswrapper[4706]: E1208 20:29:53.611153 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.187830 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm"] Dec 08 20:30:00 crc kubenswrapper[4706]: E1208 20:30:00.189113 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b14077-f0ed-4648-af39-7517847cbae5" containerName="copy" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.189133 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b14077-f0ed-4648-af39-7517847cbae5" containerName="copy" Dec 08 20:30:00 crc kubenswrapper[4706]: E1208 20:30:00.189201 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="extract-utilities" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.189210 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="extract-utilities" Dec 08 20:30:00 crc kubenswrapper[4706]: E1208 20:30:00.189234 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b14077-f0ed-4648-af39-7517847cbae5" containerName="gather" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.189242 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b14077-f0ed-4648-af39-7517847cbae5" containerName="gather" Dec 08 20:30:00 crc kubenswrapper[4706]: E1208 20:30:00.189281 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="extract-content" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.189289 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="extract-content" Dec 08 20:30:00 crc kubenswrapper[4706]: E1208 20:30:00.189301 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="registry-server" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.189335 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="registry-server" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.189577 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b14077-f0ed-4648-af39-7517847cbae5" containerName="copy" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.189599 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ec377f2-5e17-46ba-aedb-383182dd9c05" containerName="registry-server" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.189614 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b14077-f0ed-4648-af39-7517847cbae5" containerName="gather" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.190629 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.194900 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.194915 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.209535 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm"] Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.346314 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17579549-aca8-4c23-8cf5-b021fed55a7e-secret-volume\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.346433 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17579549-aca8-4c23-8cf5-b021fed55a7e-config-volume\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.346700 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nknzc\" (UniqueName: \"kubernetes.io/projected/17579549-aca8-4c23-8cf5-b021fed55a7e-kube-api-access-nknzc\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.448797 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17579549-aca8-4c23-8cf5-b021fed55a7e-secret-volume\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.448912 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17579549-aca8-4c23-8cf5-b021fed55a7e-config-volume\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.448976 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nknzc\" (UniqueName: \"kubernetes.io/projected/17579549-aca8-4c23-8cf5-b021fed55a7e-kube-api-access-nknzc\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.451555 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17579549-aca8-4c23-8cf5-b021fed55a7e-config-volume\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.604685 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17579549-aca8-4c23-8cf5-b021fed55a7e-secret-volume\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.604955 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nknzc\" (UniqueName: \"kubernetes.io/projected/17579549-aca8-4c23-8cf5-b021fed55a7e-kube-api-access-nknzc\") pod \"collect-profiles-29420430-xqbrm\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:00 crc kubenswrapper[4706]: I1208 20:30:00.824207 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:01 crc kubenswrapper[4706]: I1208 20:30:01.357587 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm"] Dec 08 20:30:02 crc kubenswrapper[4706]: I1208 20:30:02.207015 4706 generic.go:334] "Generic (PLEG): container finished" podID="17579549-aca8-4c23-8cf5-b021fed55a7e" containerID="3645acd09e40bb79456740ef288a6dc2ea828020000b61225260e5e7d001046c" exitCode=0 Dec 08 20:30:02 crc kubenswrapper[4706]: I1208 20:30:02.207111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" event={"ID":"17579549-aca8-4c23-8cf5-b021fed55a7e","Type":"ContainerDied","Data":"3645acd09e40bb79456740ef288a6dc2ea828020000b61225260e5e7d001046c"} Dec 08 20:30:02 crc kubenswrapper[4706]: I1208 20:30:02.207838 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" event={"ID":"17579549-aca8-4c23-8cf5-b021fed55a7e","Type":"ContainerStarted","Data":"61ac8e584967883b2b1497343a44d2a24a535e09e79db8ec477cc9e8bd7a55bb"} Dec 08 20:30:03 crc kubenswrapper[4706]: I1208 20:30:03.833921 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:03 crc kubenswrapper[4706]: I1208 20:30:03.963413 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17579549-aca8-4c23-8cf5-b021fed55a7e-secret-volume\") pod \"17579549-aca8-4c23-8cf5-b021fed55a7e\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " Dec 08 20:30:03 crc kubenswrapper[4706]: I1208 20:30:03.963635 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nknzc\" (UniqueName: \"kubernetes.io/projected/17579549-aca8-4c23-8cf5-b021fed55a7e-kube-api-access-nknzc\") pod \"17579549-aca8-4c23-8cf5-b021fed55a7e\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " Dec 08 20:30:03 crc kubenswrapper[4706]: I1208 20:30:03.963824 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17579549-aca8-4c23-8cf5-b021fed55a7e-config-volume\") pod \"17579549-aca8-4c23-8cf5-b021fed55a7e\" (UID: \"17579549-aca8-4c23-8cf5-b021fed55a7e\") " Dec 08 20:30:03 crc kubenswrapper[4706]: I1208 20:30:03.964862 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17579549-aca8-4c23-8cf5-b021fed55a7e-config-volume" (OuterVolumeSpecName: "config-volume") pod "17579549-aca8-4c23-8cf5-b021fed55a7e" (UID: "17579549-aca8-4c23-8cf5-b021fed55a7e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 20:30:03 crc kubenswrapper[4706]: I1208 20:30:03.971580 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17579549-aca8-4c23-8cf5-b021fed55a7e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "17579549-aca8-4c23-8cf5-b021fed55a7e" (UID: "17579549-aca8-4c23-8cf5-b021fed55a7e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 20:30:03 crc kubenswrapper[4706]: I1208 20:30:03.973773 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17579549-aca8-4c23-8cf5-b021fed55a7e-kube-api-access-nknzc" (OuterVolumeSpecName: "kube-api-access-nknzc") pod "17579549-aca8-4c23-8cf5-b021fed55a7e" (UID: "17579549-aca8-4c23-8cf5-b021fed55a7e"). InnerVolumeSpecName "kube-api-access-nknzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:30:04 crc kubenswrapper[4706]: I1208 20:30:04.067617 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17579549-aca8-4c23-8cf5-b021fed55a7e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 20:30:04 crc kubenswrapper[4706]: I1208 20:30:04.067664 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17579549-aca8-4c23-8cf5-b021fed55a7e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 20:30:04 crc kubenswrapper[4706]: I1208 20:30:04.067678 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nknzc\" (UniqueName: \"kubernetes.io/projected/17579549-aca8-4c23-8cf5-b021fed55a7e-kube-api-access-nknzc\") on node \"crc\" DevicePath \"\"" Dec 08 20:30:04 crc kubenswrapper[4706]: I1208 20:30:04.231592 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" event={"ID":"17579549-aca8-4c23-8cf5-b021fed55a7e","Type":"ContainerDied","Data":"61ac8e584967883b2b1497343a44d2a24a535e09e79db8ec477cc9e8bd7a55bb"} Dec 08 20:30:04 crc kubenswrapper[4706]: I1208 20:30:04.231643 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61ac8e584967883b2b1497343a44d2a24a535e09e79db8ec477cc9e8bd7a55bb" Dec 08 20:30:04 crc kubenswrapper[4706]: I1208 20:30:04.231700 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420430-xqbrm" Dec 08 20:30:04 crc kubenswrapper[4706]: I1208 20:30:04.917582 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm"] Dec 08 20:30:04 crc kubenswrapper[4706]: I1208 20:30:04.928708 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420385-hldgm"] Dec 08 20:30:05 crc kubenswrapper[4706]: I1208 20:30:05.609472 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:30:05 crc kubenswrapper[4706]: E1208 20:30:05.609867 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:30:05 crc kubenswrapper[4706]: I1208 20:30:05.623227 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="652e74f8-adc5-455f-91e1-f68f9ce56533" path="/var/lib/kubelet/pods/652e74f8-adc5-455f-91e1-f68f9ce56533/volumes" Dec 08 20:30:14 crc kubenswrapper[4706]: I1208 20:30:14.272976 4706 scope.go:117] "RemoveContainer" containerID="78cd172ce176672df1f8a50b3ebf190e017b13f306de53cd5e52215b247761b2" Dec 08 20:30:18 crc kubenswrapper[4706]: I1208 20:30:18.609091 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:30:18 crc kubenswrapper[4706]: E1208 20:30:18.610680 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:30:30 crc kubenswrapper[4706]: I1208 20:30:30.609392 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:30:30 crc kubenswrapper[4706]: E1208 20:30:30.610294 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:30:41 crc kubenswrapper[4706]: I1208 20:30:41.614662 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:30:41 crc kubenswrapper[4706]: E1208 20:30:41.615976 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:30:54 crc kubenswrapper[4706]: I1208 20:30:54.607922 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:30:54 crc kubenswrapper[4706]: E1208 20:30:54.610341 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:31:08 crc kubenswrapper[4706]: I1208 20:31:08.608923 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:31:08 crc kubenswrapper[4706]: I1208 20:31:08.976682 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"4d17c26bf612a2b2ecf64b164dc8622c2374233ba25c61a4ad2d539b54439cf0"} Dec 08 20:31:21 crc kubenswrapper[4706]: I1208 20:31:21.913808 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r5sc9"] Dec 08 20:31:21 crc kubenswrapper[4706]: E1208 20:31:21.916356 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17579549-aca8-4c23-8cf5-b021fed55a7e" containerName="collect-profiles" Dec 08 20:31:21 crc kubenswrapper[4706]: I1208 20:31:21.916389 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="17579549-aca8-4c23-8cf5-b021fed55a7e" containerName="collect-profiles" Dec 08 20:31:21 crc kubenswrapper[4706]: I1208 20:31:21.916696 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="17579549-aca8-4c23-8cf5-b021fed55a7e" containerName="collect-profiles" Dec 08 20:31:21 crc kubenswrapper[4706]: I1208 20:31:21.918845 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:21 crc kubenswrapper[4706]: I1208 20:31:21.932771 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5sc9"] Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.033024 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-catalog-content\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.033103 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5h2j\" (UniqueName: \"kubernetes.io/projected/bafa506b-2193-47fd-882c-f2718c4e35b2-kube-api-access-d5h2j\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.033133 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-utilities\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.135839 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-catalog-content\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.135893 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5h2j\" (UniqueName: \"kubernetes.io/projected/bafa506b-2193-47fd-882c-f2718c4e35b2-kube-api-access-d5h2j\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.135917 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-utilities\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.136492 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-catalog-content\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.136564 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-utilities\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.162445 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5h2j\" (UniqueName: \"kubernetes.io/projected/bafa506b-2193-47fd-882c-f2718c4e35b2-kube-api-access-d5h2j\") pod \"certified-operators-r5sc9\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.251097 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:22 crc kubenswrapper[4706]: I1208 20:31:22.883452 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5sc9"] Dec 08 20:31:23 crc kubenswrapper[4706]: I1208 20:31:23.133873 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5sc9" event={"ID":"bafa506b-2193-47fd-882c-f2718c4e35b2","Type":"ContainerStarted","Data":"847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c"} Dec 08 20:31:23 crc kubenswrapper[4706]: I1208 20:31:23.133933 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5sc9" event={"ID":"bafa506b-2193-47fd-882c-f2718c4e35b2","Type":"ContainerStarted","Data":"b292bc32b89658befb71feb95dd734befdac1277ed58410d7475b9813f5ed9e7"} Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.147971 4706 generic.go:334] "Generic (PLEG): container finished" podID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerID="847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c" exitCode=0 Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.148059 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5sc9" event={"ID":"bafa506b-2193-47fd-882c-f2718c4e35b2","Type":"ContainerDied","Data":"847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c"} Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.150997 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.917290 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f4gl9/must-gather-clghq"] Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.920212 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.923974 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-f4gl9"/"openshift-service-ca.crt" Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.924420 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-f4gl9"/"default-dockercfg-glmqw" Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.924586 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-f4gl9"/"kube-root-ca.crt" Dec 08 20:31:24 crc kubenswrapper[4706]: I1208 20:31:24.952454 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-f4gl9/must-gather-clghq"] Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.038955 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/74631943-0724-4b1c-a8ba-8042bdd46713-must-gather-output\") pod \"must-gather-clghq\" (UID: \"74631943-0724-4b1c-a8ba-8042bdd46713\") " pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.039081 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpp6h\" (UniqueName: \"kubernetes.io/projected/74631943-0724-4b1c-a8ba-8042bdd46713-kube-api-access-rpp6h\") pod \"must-gather-clghq\" (UID: \"74631943-0724-4b1c-a8ba-8042bdd46713\") " pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.120075 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x7z4p"] Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.123183 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.144410 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/74631943-0724-4b1c-a8ba-8042bdd46713-must-gather-output\") pod \"must-gather-clghq\" (UID: \"74631943-0724-4b1c-a8ba-8042bdd46713\") " pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.146559 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/74631943-0724-4b1c-a8ba-8042bdd46713-must-gather-output\") pod \"must-gather-clghq\" (UID: \"74631943-0724-4b1c-a8ba-8042bdd46713\") " pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.146787 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpp6h\" (UniqueName: \"kubernetes.io/projected/74631943-0724-4b1c-a8ba-8042bdd46713-kube-api-access-rpp6h\") pod \"must-gather-clghq\" (UID: \"74631943-0724-4b1c-a8ba-8042bdd46713\") " pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.163517 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7z4p"] Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.195902 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5sc9" event={"ID":"bafa506b-2193-47fd-882c-f2718c4e35b2","Type":"ContainerStarted","Data":"d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca"} Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.198528 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpp6h\" (UniqueName: \"kubernetes.io/projected/74631943-0724-4b1c-a8ba-8042bdd46713-kube-api-access-rpp6h\") pod \"must-gather-clghq\" (UID: \"74631943-0724-4b1c-a8ba-8042bdd46713\") " pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.250802 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-546pv\" (UniqueName: \"kubernetes.io/projected/35116f4b-19a4-4146-862b-33e20f44cd8a-kube-api-access-546pv\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.250930 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-utilities\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.251073 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-catalog-content\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.311639 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.353574 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-catalog-content\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.353837 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-546pv\" (UniqueName: \"kubernetes.io/projected/35116f4b-19a4-4146-862b-33e20f44cd8a-kube-api-access-546pv\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.353946 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-utilities\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.354838 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-utilities\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.354849 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-catalog-content\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.398350 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-546pv\" (UniqueName: \"kubernetes.io/projected/35116f4b-19a4-4146-862b-33e20f44cd8a-kube-api-access-546pv\") pod \"redhat-marketplace-x7z4p\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:25 crc kubenswrapper[4706]: I1208 20:31:25.451161 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:26 crc kubenswrapper[4706]: I1208 20:31:26.153976 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-f4gl9/must-gather-clghq"] Dec 08 20:31:26 crc kubenswrapper[4706]: I1208 20:31:26.317398 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7z4p"] Dec 08 20:31:27 crc kubenswrapper[4706]: I1208 20:31:27.225165 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/must-gather-clghq" event={"ID":"74631943-0724-4b1c-a8ba-8042bdd46713","Type":"ContainerStarted","Data":"ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23"} Dec 08 20:31:27 crc kubenswrapper[4706]: I1208 20:31:27.225911 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/must-gather-clghq" event={"ID":"74631943-0724-4b1c-a8ba-8042bdd46713","Type":"ContainerStarted","Data":"d99100593e33d22abe033436f7a8b7b993fb2d49f0870c60eb5b40f668b155b4"} Dec 08 20:31:27 crc kubenswrapper[4706]: I1208 20:31:27.227559 4706 generic.go:334] "Generic (PLEG): container finished" podID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerID="d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca" exitCode=0 Dec 08 20:31:27 crc kubenswrapper[4706]: I1208 20:31:27.227655 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5sc9" event={"ID":"bafa506b-2193-47fd-882c-f2718c4e35b2","Type":"ContainerDied","Data":"d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca"} Dec 08 20:31:27 crc kubenswrapper[4706]: I1208 20:31:27.229547 4706 generic.go:334] "Generic (PLEG): container finished" podID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerID="27da3651a096aa0db9f4487143bd54b8a89b1ec0c56595ad63290331bf09d7d0" exitCode=0 Dec 08 20:31:27 crc kubenswrapper[4706]: I1208 20:31:27.229583 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7z4p" event={"ID":"35116f4b-19a4-4146-862b-33e20f44cd8a","Type":"ContainerDied","Data":"27da3651a096aa0db9f4487143bd54b8a89b1ec0c56595ad63290331bf09d7d0"} Dec 08 20:31:27 crc kubenswrapper[4706]: I1208 20:31:27.229616 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7z4p" event={"ID":"35116f4b-19a4-4146-862b-33e20f44cd8a","Type":"ContainerStarted","Data":"664b0f01a1280eef49b065ee4cb7f0a31346f57a8d7bd811e3c84d63288a1a85"} Dec 08 20:31:28 crc kubenswrapper[4706]: I1208 20:31:28.246114 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/must-gather-clghq" event={"ID":"74631943-0724-4b1c-a8ba-8042bdd46713","Type":"ContainerStarted","Data":"c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2"} Dec 08 20:31:28 crc kubenswrapper[4706]: I1208 20:31:28.249651 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5sc9" event={"ID":"bafa506b-2193-47fd-882c-f2718c4e35b2","Type":"ContainerStarted","Data":"6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4"} Dec 08 20:31:28 crc kubenswrapper[4706]: I1208 20:31:28.253002 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7z4p" event={"ID":"35116f4b-19a4-4146-862b-33e20f44cd8a","Type":"ContainerStarted","Data":"1967f9364e6fc5cd18267b227b7f377f7ffb5614964877b3c0b8035c65706327"} Dec 08 20:31:28 crc kubenswrapper[4706]: I1208 20:31:28.294330 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-f4gl9/must-gather-clghq" podStartSLOduration=4.294299383 podStartE2EDuration="4.294299383s" podCreationTimestamp="2025-12-08 20:31:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 20:31:28.276171468 +0000 UTC m=+4170.918372481" watchObservedRunningTime="2025-12-08 20:31:28.294299383 +0000 UTC m=+4170.936500386" Dec 08 20:31:28 crc kubenswrapper[4706]: I1208 20:31:28.358237 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r5sc9" podStartSLOduration=3.7331617489999998 podStartE2EDuration="7.358199508s" podCreationTimestamp="2025-12-08 20:31:21 +0000 UTC" firstStartedPulling="2025-12-08 20:31:24.150706488 +0000 UTC m=+4166.792907491" lastFinishedPulling="2025-12-08 20:31:27.775744247 +0000 UTC m=+4170.417945250" observedRunningTime="2025-12-08 20:31:28.346423933 +0000 UTC m=+4170.988624946" watchObservedRunningTime="2025-12-08 20:31:28.358199508 +0000 UTC m=+4171.000400511" Dec 08 20:31:29 crc kubenswrapper[4706]: I1208 20:31:29.268148 4706 generic.go:334] "Generic (PLEG): container finished" podID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerID="1967f9364e6fc5cd18267b227b7f377f7ffb5614964877b3c0b8035c65706327" exitCode=0 Dec 08 20:31:29 crc kubenswrapper[4706]: I1208 20:31:29.268296 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7z4p" event={"ID":"35116f4b-19a4-4146-862b-33e20f44cd8a","Type":"ContainerDied","Data":"1967f9364e6fc5cd18267b227b7f377f7ffb5614964877b3c0b8035c65706327"} Dec 08 20:31:30 crc kubenswrapper[4706]: I1208 20:31:30.288603 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7z4p" event={"ID":"35116f4b-19a4-4146-862b-33e20f44cd8a","Type":"ContainerStarted","Data":"3921985fcf2d2b6f044491e4995bf0bccdfba1377f48fd21d8edf24c7bbd21af"} Dec 08 20:31:30 crc kubenswrapper[4706]: I1208 20:31:30.316945 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x7z4p" podStartSLOduration=2.829135693 podStartE2EDuration="5.316911648s" podCreationTimestamp="2025-12-08 20:31:25 +0000 UTC" firstStartedPulling="2025-12-08 20:31:27.231336538 +0000 UTC m=+4169.873537541" lastFinishedPulling="2025-12-08 20:31:29.719112473 +0000 UTC m=+4172.361313496" observedRunningTime="2025-12-08 20:31:30.311770732 +0000 UTC m=+4172.953971735" watchObservedRunningTime="2025-12-08 20:31:30.316911648 +0000 UTC m=+4172.959112651" Dec 08 20:31:31 crc kubenswrapper[4706]: I1208 20:31:31.873963 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-jpzz9"] Dec 08 20:31:31 crc kubenswrapper[4706]: I1208 20:31:31.890198 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.027242 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cmhr\" (UniqueName: \"kubernetes.io/projected/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-kube-api-access-6cmhr\") pod \"crc-debug-jpzz9\" (UID: \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\") " pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.027679 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-host\") pod \"crc-debug-jpzz9\" (UID: \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\") " pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.131283 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cmhr\" (UniqueName: \"kubernetes.io/projected/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-kube-api-access-6cmhr\") pod \"crc-debug-jpzz9\" (UID: \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\") " pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.131458 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-host\") pod \"crc-debug-jpzz9\" (UID: \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\") " pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.131607 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-host\") pod \"crc-debug-jpzz9\" (UID: \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\") " pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.161017 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cmhr\" (UniqueName: \"kubernetes.io/projected/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-kube-api-access-6cmhr\") pod \"crc-debug-jpzz9\" (UID: \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\") " pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.233294 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.251698 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.252120 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:32 crc kubenswrapper[4706]: W1208 20:31:32.304505 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f9362e6_5a7c_42ce_bab0_0417cf3f390f.slice/crio-36efa56926ef2af9bee3bcd3aeba28f55042829b9236f9cf8657eb9c7b430055 WatchSource:0}: Error finding container 36efa56926ef2af9bee3bcd3aeba28f55042829b9236f9cf8657eb9c7b430055: Status 404 returned error can't find the container with id 36efa56926ef2af9bee3bcd3aeba28f55042829b9236f9cf8657eb9c7b430055 Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.383119 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:32 crc kubenswrapper[4706]: I1208 20:31:32.452065 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:33 crc kubenswrapper[4706]: I1208 20:31:33.327789 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" event={"ID":"1f9362e6-5a7c-42ce-bab0-0417cf3f390f","Type":"ContainerStarted","Data":"8a460d97857db14c7408e3c235fcb29ec2d14109b90a2b2489e6a4bd8da8062b"} Dec 08 20:31:33 crc kubenswrapper[4706]: I1208 20:31:33.328712 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" event={"ID":"1f9362e6-5a7c-42ce-bab0-0417cf3f390f","Type":"ContainerStarted","Data":"36efa56926ef2af9bee3bcd3aeba28f55042829b9236f9cf8657eb9c7b430055"} Dec 08 20:31:33 crc kubenswrapper[4706]: I1208 20:31:33.361544 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" podStartSLOduration=2.361522416 podStartE2EDuration="2.361522416s" podCreationTimestamp="2025-12-08 20:31:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 20:31:33.346779627 +0000 UTC m=+4175.988980640" watchObservedRunningTime="2025-12-08 20:31:33.361522416 +0000 UTC m=+4176.003723419" Dec 08 20:31:33 crc kubenswrapper[4706]: I1208 20:31:33.712851 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5sc9"] Dec 08 20:31:34 crc kubenswrapper[4706]: I1208 20:31:34.350298 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r5sc9" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerName="registry-server" containerID="cri-o://6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4" gracePeriod=2 Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.018875 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.122204 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-catalog-content\") pod \"bafa506b-2193-47fd-882c-f2718c4e35b2\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.122526 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-utilities\") pod \"bafa506b-2193-47fd-882c-f2718c4e35b2\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.122581 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5h2j\" (UniqueName: \"kubernetes.io/projected/bafa506b-2193-47fd-882c-f2718c4e35b2-kube-api-access-d5h2j\") pod \"bafa506b-2193-47fd-882c-f2718c4e35b2\" (UID: \"bafa506b-2193-47fd-882c-f2718c4e35b2\") " Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.124387 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-utilities" (OuterVolumeSpecName: "utilities") pod "bafa506b-2193-47fd-882c-f2718c4e35b2" (UID: "bafa506b-2193-47fd-882c-f2718c4e35b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.133391 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bafa506b-2193-47fd-882c-f2718c4e35b2-kube-api-access-d5h2j" (OuterVolumeSpecName: "kube-api-access-d5h2j") pod "bafa506b-2193-47fd-882c-f2718c4e35b2" (UID: "bafa506b-2193-47fd-882c-f2718c4e35b2"). InnerVolumeSpecName "kube-api-access-d5h2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.208031 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bafa506b-2193-47fd-882c-f2718c4e35b2" (UID: "bafa506b-2193-47fd-882c-f2718c4e35b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.225733 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.225803 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5h2j\" (UniqueName: \"kubernetes.io/projected/bafa506b-2193-47fd-882c-f2718c4e35b2-kube-api-access-d5h2j\") on node \"crc\" DevicePath \"\"" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.225818 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bafa506b-2193-47fd-882c-f2718c4e35b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.368096 4706 generic.go:334] "Generic (PLEG): container finished" podID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerID="6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4" exitCode=0 Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.368145 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5sc9" event={"ID":"bafa506b-2193-47fd-882c-f2718c4e35b2","Type":"ContainerDied","Data":"6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4"} Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.368175 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5sc9" event={"ID":"bafa506b-2193-47fd-882c-f2718c4e35b2","Type":"ContainerDied","Data":"b292bc32b89658befb71feb95dd734befdac1277ed58410d7475b9813f5ed9e7"} Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.368195 4706 scope.go:117] "RemoveContainer" containerID="6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.368369 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5sc9" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.423353 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5sc9"] Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.431774 4706 scope.go:117] "RemoveContainer" containerID="d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.446554 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r5sc9"] Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.452750 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.454484 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.495502 4706 scope.go:117] "RemoveContainer" containerID="847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.554585 4706 scope.go:117] "RemoveContainer" containerID="6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4" Dec 08 20:31:35 crc kubenswrapper[4706]: E1208 20:31:35.557897 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4\": container with ID starting with 6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4 not found: ID does not exist" containerID="6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.557944 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4"} err="failed to get container status \"6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4\": rpc error: code = NotFound desc = could not find container \"6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4\": container with ID starting with 6924321ced82899dacd546acf0271c0295896956f4cc6e2ede506b966a48d4b4 not found: ID does not exist" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.557975 4706 scope.go:117] "RemoveContainer" containerID="d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca" Dec 08 20:31:35 crc kubenswrapper[4706]: E1208 20:31:35.558977 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca\": container with ID starting with d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca not found: ID does not exist" containerID="d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.559002 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca"} err="failed to get container status \"d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca\": rpc error: code = NotFound desc = could not find container \"d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca\": container with ID starting with d36e3088f057cc354a8fe01535535fce47320808194d942ed3bb2e15b38881ca not found: ID does not exist" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.559019 4706 scope.go:117] "RemoveContainer" containerID="847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c" Dec 08 20:31:35 crc kubenswrapper[4706]: E1208 20:31:35.559478 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c\": container with ID starting with 847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c not found: ID does not exist" containerID="847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.559512 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c"} err="failed to get container status \"847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c\": rpc error: code = NotFound desc = could not find container \"847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c\": container with ID starting with 847ba1e6166be6126a40a37a4def8543b9b071b31ac9f97712916e903a5c4f2c not found: ID does not exist" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.563283 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:35 crc kubenswrapper[4706]: I1208 20:31:35.631675 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" path="/var/lib/kubelet/pods/bafa506b-2193-47fd-882c-f2718c4e35b2/volumes" Dec 08 20:31:36 crc kubenswrapper[4706]: I1208 20:31:36.441550 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:37 crc kubenswrapper[4706]: I1208 20:31:37.899153 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7z4p"] Dec 08 20:31:39 crc kubenswrapper[4706]: I1208 20:31:39.415231 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x7z4p" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerName="registry-server" containerID="cri-o://3921985fcf2d2b6f044491e4995bf0bccdfba1377f48fd21d8edf24c7bbd21af" gracePeriod=2 Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.488931 4706 generic.go:334] "Generic (PLEG): container finished" podID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerID="3921985fcf2d2b6f044491e4995bf0bccdfba1377f48fd21d8edf24c7bbd21af" exitCode=0 Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.489646 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7z4p" event={"ID":"35116f4b-19a4-4146-862b-33e20f44cd8a","Type":"ContainerDied","Data":"3921985fcf2d2b6f044491e4995bf0bccdfba1377f48fd21d8edf24c7bbd21af"} Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.833784 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.899364 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-utilities\") pod \"35116f4b-19a4-4146-862b-33e20f44cd8a\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.899580 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-546pv\" (UniqueName: \"kubernetes.io/projected/35116f4b-19a4-4146-862b-33e20f44cd8a-kube-api-access-546pv\") pod \"35116f4b-19a4-4146-862b-33e20f44cd8a\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.900002 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-catalog-content\") pod \"35116f4b-19a4-4146-862b-33e20f44cd8a\" (UID: \"35116f4b-19a4-4146-862b-33e20f44cd8a\") " Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.900552 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-utilities" (OuterVolumeSpecName: "utilities") pod "35116f4b-19a4-4146-862b-33e20f44cd8a" (UID: "35116f4b-19a4-4146-862b-33e20f44cd8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.900966 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.928184 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35116f4b-19a4-4146-862b-33e20f44cd8a-kube-api-access-546pv" (OuterVolumeSpecName: "kube-api-access-546pv") pod "35116f4b-19a4-4146-862b-33e20f44cd8a" (UID: "35116f4b-19a4-4146-862b-33e20f44cd8a"). InnerVolumeSpecName "kube-api-access-546pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:31:40 crc kubenswrapper[4706]: I1208 20:31:40.971403 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35116f4b-19a4-4146-862b-33e20f44cd8a" (UID: "35116f4b-19a4-4146-862b-33e20f44cd8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:31:41 crc kubenswrapper[4706]: I1208 20:31:41.002893 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-546pv\" (UniqueName: \"kubernetes.io/projected/35116f4b-19a4-4146-862b-33e20f44cd8a-kube-api-access-546pv\") on node \"crc\" DevicePath \"\"" Dec 08 20:31:41 crc kubenswrapper[4706]: I1208 20:31:41.002945 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35116f4b-19a4-4146-862b-33e20f44cd8a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:31:41 crc kubenswrapper[4706]: I1208 20:31:41.504437 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7z4p" event={"ID":"35116f4b-19a4-4146-862b-33e20f44cd8a","Type":"ContainerDied","Data":"664b0f01a1280eef49b065ee4cb7f0a31346f57a8d7bd811e3c84d63288a1a85"} Dec 08 20:31:41 crc kubenswrapper[4706]: I1208 20:31:41.504485 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:31:41 crc kubenswrapper[4706]: I1208 20:31:41.504814 4706 scope.go:117] "RemoveContainer" containerID="3921985fcf2d2b6f044491e4995bf0bccdfba1377f48fd21d8edf24c7bbd21af" Dec 08 20:31:42 crc kubenswrapper[4706]: I1208 20:31:42.948622 4706 scope.go:117] "RemoveContainer" containerID="1967f9364e6fc5cd18267b227b7f377f7ffb5614964877b3c0b8035c65706327" Dec 08 20:31:42 crc kubenswrapper[4706]: I1208 20:31:42.985812 4706 scope.go:117] "RemoveContainer" containerID="27da3651a096aa0db9f4487143bd54b8a89b1ec0c56595ad63290331bf09d7d0" Dec 08 20:32:12 crc kubenswrapper[4706]: I1208 20:32:12.938496 4706 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod35116f4b-19a4-4146-862b-33e20f44cd8a"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod35116f4b-19a4-4146-862b-33e20f44cd8a] : Timed out while waiting for systemd to remove kubepods-burstable-pod35116f4b_19a4_4146_862b_33e20f44cd8a.slice" Dec 08 20:32:12 crc kubenswrapper[4706]: E1208 20:32:12.939458 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable pod35116f4b-19a4-4146-862b-33e20f44cd8a] : unable to destroy cgroup paths for cgroup [kubepods burstable pod35116f4b-19a4-4146-862b-33e20f44cd8a] : Timed out while waiting for systemd to remove kubepods-burstable-pod35116f4b_19a4_4146_862b_33e20f44cd8a.slice" pod="openshift-marketplace/redhat-marketplace-x7z4p" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" Dec 08 20:32:13 crc kubenswrapper[4706]: I1208 20:32:13.922460 4706 generic.go:334] "Generic (PLEG): container finished" podID="1f9362e6-5a7c-42ce-bab0-0417cf3f390f" containerID="8a460d97857db14c7408e3c235fcb29ec2d14109b90a2b2489e6a4bd8da8062b" exitCode=0 Dec 08 20:32:13 crc kubenswrapper[4706]: I1208 20:32:13.923494 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7z4p" Dec 08 20:32:13 crc kubenswrapper[4706]: I1208 20:32:13.922628 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" event={"ID":"1f9362e6-5a7c-42ce-bab0-0417cf3f390f","Type":"ContainerDied","Data":"8a460d97857db14c7408e3c235fcb29ec2d14109b90a2b2489e6a4bd8da8062b"} Dec 08 20:32:13 crc kubenswrapper[4706]: I1208 20:32:13.985729 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7z4p"] Dec 08 20:32:14 crc kubenswrapper[4706]: I1208 20:32:14.002889 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7z4p"] Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.084404 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.129321 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-jpzz9"] Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.141731 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-jpzz9"] Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.195631 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cmhr\" (UniqueName: \"kubernetes.io/projected/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-kube-api-access-6cmhr\") pod \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\" (UID: \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\") " Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.195773 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-host\") pod \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\" (UID: \"1f9362e6-5a7c-42ce-bab0-0417cf3f390f\") " Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.196154 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-host" (OuterVolumeSpecName: "host") pod "1f9362e6-5a7c-42ce-bab0-0417cf3f390f" (UID: "1f9362e6-5a7c-42ce-bab0-0417cf3f390f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.196577 4706 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-host\") on node \"crc\" DevicePath \"\"" Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.215436 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-kube-api-access-6cmhr" (OuterVolumeSpecName: "kube-api-access-6cmhr") pod "1f9362e6-5a7c-42ce-bab0-0417cf3f390f" (UID: "1f9362e6-5a7c-42ce-bab0-0417cf3f390f"). InnerVolumeSpecName "kube-api-access-6cmhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.298812 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cmhr\" (UniqueName: \"kubernetes.io/projected/1f9362e6-5a7c-42ce-bab0-0417cf3f390f-kube-api-access-6cmhr\") on node \"crc\" DevicePath \"\"" Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.624846 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9362e6-5a7c-42ce-bab0-0417cf3f390f" path="/var/lib/kubelet/pods/1f9362e6-5a7c-42ce-bab0-0417cf3f390f/volumes" Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.625596 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" path="/var/lib/kubelet/pods/35116f4b-19a4-4146-862b-33e20f44cd8a/volumes" Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.946577 4706 scope.go:117] "RemoveContainer" containerID="8a460d97857db14c7408e3c235fcb29ec2d14109b90a2b2489e6a4bd8da8062b" Dec 08 20:32:15 crc kubenswrapper[4706]: I1208 20:32:15.946615 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-jpzz9" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.388722 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-b45jc"] Dec 08 20:32:16 crc kubenswrapper[4706]: E1208 20:32:16.389820 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerName="extract-content" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.389841 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerName="extract-content" Dec 08 20:32:16 crc kubenswrapper[4706]: E1208 20:32:16.389867 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerName="registry-server" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.389875 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerName="registry-server" Dec 08 20:32:16 crc kubenswrapper[4706]: E1208 20:32:16.389888 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9362e6-5a7c-42ce-bab0-0417cf3f390f" containerName="container-00" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.389895 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9362e6-5a7c-42ce-bab0-0417cf3f390f" containerName="container-00" Dec 08 20:32:16 crc kubenswrapper[4706]: E1208 20:32:16.389904 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerName="extract-utilities" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.389912 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerName="extract-utilities" Dec 08 20:32:16 crc kubenswrapper[4706]: E1208 20:32:16.389947 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerName="extract-content" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.389954 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerName="extract-content" Dec 08 20:32:16 crc kubenswrapper[4706]: E1208 20:32:16.389971 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerName="registry-server" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.389978 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerName="registry-server" Dec 08 20:32:16 crc kubenswrapper[4706]: E1208 20:32:16.390001 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerName="extract-utilities" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.390008 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerName="extract-utilities" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.390326 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9362e6-5a7c-42ce-bab0-0417cf3f390f" containerName="container-00" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.390352 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="35116f4b-19a4-4146-862b-33e20f44cd8a" containerName="registry-server" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.390377 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bafa506b-2193-47fd-882c-f2718c4e35b2" containerName="registry-server" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.391559 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.529248 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2m9h\" (UniqueName: \"kubernetes.io/projected/e0948539-2ffd-4491-bdbb-5b402220a7f5-kube-api-access-k2m9h\") pod \"crc-debug-b45jc\" (UID: \"e0948539-2ffd-4491-bdbb-5b402220a7f5\") " pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.529800 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0948539-2ffd-4491-bdbb-5b402220a7f5-host\") pod \"crc-debug-b45jc\" (UID: \"e0948539-2ffd-4491-bdbb-5b402220a7f5\") " pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.633123 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0948539-2ffd-4491-bdbb-5b402220a7f5-host\") pod \"crc-debug-b45jc\" (UID: \"e0948539-2ffd-4491-bdbb-5b402220a7f5\") " pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.633346 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2m9h\" (UniqueName: \"kubernetes.io/projected/e0948539-2ffd-4491-bdbb-5b402220a7f5-kube-api-access-k2m9h\") pod \"crc-debug-b45jc\" (UID: \"e0948539-2ffd-4491-bdbb-5b402220a7f5\") " pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:16 crc kubenswrapper[4706]: I1208 20:32:16.633878 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0948539-2ffd-4491-bdbb-5b402220a7f5-host\") pod \"crc-debug-b45jc\" (UID: \"e0948539-2ffd-4491-bdbb-5b402220a7f5\") " pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:17 crc kubenswrapper[4706]: I1208 20:32:17.304633 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2m9h\" (UniqueName: \"kubernetes.io/projected/e0948539-2ffd-4491-bdbb-5b402220a7f5-kube-api-access-k2m9h\") pod \"crc-debug-b45jc\" (UID: \"e0948539-2ffd-4491-bdbb-5b402220a7f5\") " pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:17 crc kubenswrapper[4706]: I1208 20:32:17.311799 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:17 crc kubenswrapper[4706]: I1208 20:32:17.973216 4706 generic.go:334] "Generic (PLEG): container finished" podID="e0948539-2ffd-4491-bdbb-5b402220a7f5" containerID="6baf48c85b9958a01fd14f95931a1c0de0f86476b22e354d63b0291f52d33c5d" exitCode=0 Dec 08 20:32:17 crc kubenswrapper[4706]: I1208 20:32:17.973307 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/crc-debug-b45jc" event={"ID":"e0948539-2ffd-4491-bdbb-5b402220a7f5","Type":"ContainerDied","Data":"6baf48c85b9958a01fd14f95931a1c0de0f86476b22e354d63b0291f52d33c5d"} Dec 08 20:32:17 crc kubenswrapper[4706]: I1208 20:32:17.973837 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/crc-debug-b45jc" event={"ID":"e0948539-2ffd-4491-bdbb-5b402220a7f5","Type":"ContainerStarted","Data":"8c92b93b9b9f5476a9b57bca5f39e1c94d2e0a61f5ae85b083ca13cda2994d7b"} Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.108500 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.136454 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0948539-2ffd-4491-bdbb-5b402220a7f5-host\") pod \"e0948539-2ffd-4491-bdbb-5b402220a7f5\" (UID: \"e0948539-2ffd-4491-bdbb-5b402220a7f5\") " Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.136635 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0948539-2ffd-4491-bdbb-5b402220a7f5-host" (OuterVolumeSpecName: "host") pod "e0948539-2ffd-4491-bdbb-5b402220a7f5" (UID: "e0948539-2ffd-4491-bdbb-5b402220a7f5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.136677 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2m9h\" (UniqueName: \"kubernetes.io/projected/e0948539-2ffd-4491-bdbb-5b402220a7f5-kube-api-access-k2m9h\") pod \"e0948539-2ffd-4491-bdbb-5b402220a7f5\" (UID: \"e0948539-2ffd-4491-bdbb-5b402220a7f5\") " Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.137545 4706 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0948539-2ffd-4491-bdbb-5b402220a7f5-host\") on node \"crc\" DevicePath \"\"" Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.162756 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0948539-2ffd-4491-bdbb-5b402220a7f5-kube-api-access-k2m9h" (OuterVolumeSpecName: "kube-api-access-k2m9h") pod "e0948539-2ffd-4491-bdbb-5b402220a7f5" (UID: "e0948539-2ffd-4491-bdbb-5b402220a7f5"). InnerVolumeSpecName "kube-api-access-k2m9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.240488 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2m9h\" (UniqueName: \"kubernetes.io/projected/e0948539-2ffd-4491-bdbb-5b402220a7f5-kube-api-access-k2m9h\") on node \"crc\" DevicePath \"\"" Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.287743 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-b45jc"] Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.299041 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-b45jc"] Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.623726 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0948539-2ffd-4491-bdbb-5b402220a7f5" path="/var/lib/kubelet/pods/e0948539-2ffd-4491-bdbb-5b402220a7f5/volumes" Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.996753 4706 scope.go:117] "RemoveContainer" containerID="6baf48c85b9958a01fd14f95931a1c0de0f86476b22e354d63b0291f52d33c5d" Dec 08 20:32:19 crc kubenswrapper[4706]: I1208 20:32:19.996774 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-b45jc" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.616530 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-52t7p"] Dec 08 20:32:20 crc kubenswrapper[4706]: E1208 20:32:20.617155 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0948539-2ffd-4491-bdbb-5b402220a7f5" containerName="container-00" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.617181 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0948539-2ffd-4491-bdbb-5b402220a7f5" containerName="container-00" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.617509 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0948539-2ffd-4491-bdbb-5b402220a7f5" containerName="container-00" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.618830 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.676937 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4206f9d-5fae-4379-9efa-b5380843d090-host\") pod \"crc-debug-52t7p\" (UID: \"e4206f9d-5fae-4379-9efa-b5380843d090\") " pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.676994 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjwc6\" (UniqueName: \"kubernetes.io/projected/e4206f9d-5fae-4379-9efa-b5380843d090-kube-api-access-cjwc6\") pod \"crc-debug-52t7p\" (UID: \"e4206f9d-5fae-4379-9efa-b5380843d090\") " pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.779417 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4206f9d-5fae-4379-9efa-b5380843d090-host\") pod \"crc-debug-52t7p\" (UID: \"e4206f9d-5fae-4379-9efa-b5380843d090\") " pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.779805 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjwc6\" (UniqueName: \"kubernetes.io/projected/e4206f9d-5fae-4379-9efa-b5380843d090-kube-api-access-cjwc6\") pod \"crc-debug-52t7p\" (UID: \"e4206f9d-5fae-4379-9efa-b5380843d090\") " pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.779587 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4206f9d-5fae-4379-9efa-b5380843d090-host\") pod \"crc-debug-52t7p\" (UID: \"e4206f9d-5fae-4379-9efa-b5380843d090\") " pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.813468 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjwc6\" (UniqueName: \"kubernetes.io/projected/e4206f9d-5fae-4379-9efa-b5380843d090-kube-api-access-cjwc6\") pod \"crc-debug-52t7p\" (UID: \"e4206f9d-5fae-4379-9efa-b5380843d090\") " pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:20 crc kubenswrapper[4706]: I1208 20:32:20.945134 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:21 crc kubenswrapper[4706]: I1208 20:32:21.013208 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/crc-debug-52t7p" event={"ID":"e4206f9d-5fae-4379-9efa-b5380843d090","Type":"ContainerStarted","Data":"44fd164805588e22a430523b44b7131f381aba4d0dc141ef75fdbaf3d5bd8815"} Dec 08 20:32:22 crc kubenswrapper[4706]: I1208 20:32:22.031302 4706 generic.go:334] "Generic (PLEG): container finished" podID="e4206f9d-5fae-4379-9efa-b5380843d090" containerID="0ced629694e8333fb01955db762b184d6ec8bcf69146d8db907d1fb4222dc5d3" exitCode=0 Dec 08 20:32:22 crc kubenswrapper[4706]: I1208 20:32:22.031443 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/crc-debug-52t7p" event={"ID":"e4206f9d-5fae-4379-9efa-b5380843d090","Type":"ContainerDied","Data":"0ced629694e8333fb01955db762b184d6ec8bcf69146d8db907d1fb4222dc5d3"} Dec 08 20:32:22 crc kubenswrapper[4706]: I1208 20:32:22.112890 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-52t7p"] Dec 08 20:32:22 crc kubenswrapper[4706]: I1208 20:32:22.138745 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f4gl9/crc-debug-52t7p"] Dec 08 20:32:23 crc kubenswrapper[4706]: I1208 20:32:23.177146 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:32:23 crc kubenswrapper[4706]: I1208 20:32:23.346964 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4206f9d-5fae-4379-9efa-b5380843d090-host\") pod \"e4206f9d-5fae-4379-9efa-b5380843d090\" (UID: \"e4206f9d-5fae-4379-9efa-b5380843d090\") " Dec 08 20:32:23 crc kubenswrapper[4706]: I1208 20:32:23.347071 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4206f9d-5fae-4379-9efa-b5380843d090-host" (OuterVolumeSpecName: "host") pod "e4206f9d-5fae-4379-9efa-b5380843d090" (UID: "e4206f9d-5fae-4379-9efa-b5380843d090"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 20:32:23 crc kubenswrapper[4706]: I1208 20:32:23.347139 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjwc6\" (UniqueName: \"kubernetes.io/projected/e4206f9d-5fae-4379-9efa-b5380843d090-kube-api-access-cjwc6\") pod \"e4206f9d-5fae-4379-9efa-b5380843d090\" (UID: \"e4206f9d-5fae-4379-9efa-b5380843d090\") " Dec 08 20:32:23 crc kubenswrapper[4706]: I1208 20:32:23.348458 4706 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4206f9d-5fae-4379-9efa-b5380843d090-host\") on node \"crc\" DevicePath \"\"" Dec 08 20:32:23 crc kubenswrapper[4706]: I1208 20:32:23.354980 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4206f9d-5fae-4379-9efa-b5380843d090-kube-api-access-cjwc6" (OuterVolumeSpecName: "kube-api-access-cjwc6") pod "e4206f9d-5fae-4379-9efa-b5380843d090" (UID: "e4206f9d-5fae-4379-9efa-b5380843d090"). InnerVolumeSpecName "kube-api-access-cjwc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:32:23 crc kubenswrapper[4706]: I1208 20:32:23.450583 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjwc6\" (UniqueName: \"kubernetes.io/projected/e4206f9d-5fae-4379-9efa-b5380843d090-kube-api-access-cjwc6\") on node \"crc\" DevicePath \"\"" Dec 08 20:32:23 crc kubenswrapper[4706]: I1208 20:32:23.626241 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4206f9d-5fae-4379-9efa-b5380843d090" path="/var/lib/kubelet/pods/e4206f9d-5fae-4379-9efa-b5380843d090/volumes" Dec 08 20:32:24 crc kubenswrapper[4706]: I1208 20:32:24.056676 4706 scope.go:117] "RemoveContainer" containerID="0ced629694e8333fb01955db762b184d6ec8bcf69146d8db907d1fb4222dc5d3" Dec 08 20:32:24 crc kubenswrapper[4706]: I1208 20:32:24.056731 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/crc-debug-52t7p" Dec 08 20:33:18 crc kubenswrapper[4706]: I1208 20:33:18.151791 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88233d80-672d-47c3-9a26-70c22f748e29/init-config-reloader/0.log" Dec 08 20:33:18 crc kubenswrapper[4706]: I1208 20:33:18.402933 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88233d80-672d-47c3-9a26-70c22f748e29/config-reloader/0.log" Dec 08 20:33:18 crc kubenswrapper[4706]: I1208 20:33:18.482961 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88233d80-672d-47c3-9a26-70c22f748e29/alertmanager/0.log" Dec 08 20:33:18 crc kubenswrapper[4706]: I1208 20:33:18.507756 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88233d80-672d-47c3-9a26-70c22f748e29/init-config-reloader/0.log" Dec 08 20:33:18 crc kubenswrapper[4706]: I1208 20:33:18.703016 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8466877bdb-zbttb_b0acc09f-0610-4121-b42b-ebe7274b5c5d/barbican-api/0.log" Dec 08 20:33:18 crc kubenswrapper[4706]: I1208 20:33:18.734541 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8466877bdb-zbttb_b0acc09f-0610-4121-b42b-ebe7274b5c5d/barbican-api-log/0.log" Dec 08 20:33:18 crc kubenswrapper[4706]: I1208 20:33:18.820238 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-788468f6bb-gm7rb_aa3a5522-ce55-49b7-a575-3f83c373b2ac/barbican-keystone-listener/0.log" Dec 08 20:33:19 crc kubenswrapper[4706]: I1208 20:33:19.309071 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b989c8b75-k95kd_2cc4f4da-55de-4f4e-9e3e-099e148987a2/barbican-worker-log/0.log" Dec 08 20:33:19 crc kubenswrapper[4706]: I1208 20:33:19.355562 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-788468f6bb-gm7rb_aa3a5522-ce55-49b7-a575-3f83c373b2ac/barbican-keystone-listener-log/0.log" Dec 08 20:33:19 crc kubenswrapper[4706]: I1208 20:33:19.412716 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b989c8b75-k95kd_2cc4f4da-55de-4f4e-9e3e-099e148987a2/barbican-worker/0.log" Dec 08 20:33:19 crc kubenswrapper[4706]: I1208 20:33:19.624671 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-267vl_91cd912d-3492-4709-99e9-94fa496bde14/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:19 crc kubenswrapper[4706]: I1208 20:33:19.827287 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42/ceilometer-central-agent/0.log" Dec 08 20:33:19 crc kubenswrapper[4706]: I1208 20:33:19.954558 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42/ceilometer-notification-agent/0.log" Dec 08 20:33:19 crc kubenswrapper[4706]: I1208 20:33:19.974480 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42/proxy-httpd/0.log" Dec 08 20:33:20 crc kubenswrapper[4706]: I1208 20:33:20.020710 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_e15bf8d7-2b7a-4e95-acf5-c9f32bc15b42/sg-core/0.log" Dec 08 20:33:20 crc kubenswrapper[4706]: I1208 20:33:20.175187 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_81a640d0-b89a-49be-aa17-00ad9d0d444a/cinder-api-log/0.log" Dec 08 20:33:20 crc kubenswrapper[4706]: I1208 20:33:20.271911 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_81a640d0-b89a-49be-aa17-00ad9d0d444a/cinder-api/0.log" Dec 08 20:33:20 crc kubenswrapper[4706]: I1208 20:33:20.475027 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_33d60263-f73a-470d-b0de-60eb2a8ecfd5/cinder-scheduler/0.log" Dec 08 20:33:20 crc kubenswrapper[4706]: I1208 20:33:20.541671 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_33d60263-f73a-470d-b0de-60eb2a8ecfd5/probe/0.log" Dec 08 20:33:20 crc kubenswrapper[4706]: I1208 20:33:20.710981 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_b0316b44-87fb-44dd-890c-dace6fb0ab3e/cloudkitty-api/0.log" Dec 08 20:33:20 crc kubenswrapper[4706]: I1208 20:33:20.826590 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_b0316b44-87fb-44dd-890c-dace6fb0ab3e/cloudkitty-api-log/0.log" Dec 08 20:33:20 crc kubenswrapper[4706]: I1208 20:33:20.907808 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_a7e3aaf0-5179-4837-b6ca-53276d637d10/loki-compactor/0.log" Dec 08 20:33:21 crc kubenswrapper[4706]: I1208 20:33:21.147837 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-664b687b54-zfbc5_081f2271-cba5-43e4-a179-7fd4389c6531/loki-distributor/0.log" Dec 08 20:33:21 crc kubenswrapper[4706]: I1208 20:33:21.203197 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-bc75944f-7tk8c_d9d23ed1-9384-4bf8-adeb-2bc3340279ed/gateway/0.log" Dec 08 20:33:21 crc kubenswrapper[4706]: I1208 20:33:21.451718 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-bc75944f-m4nfc_7811fdf0-d557-47cb-b700-4a072602f8c7/gateway/0.log" Dec 08 20:33:21 crc kubenswrapper[4706]: I1208 20:33:21.525755 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_4d1845b4-bbc3-4661-83d4-72b972fa288a/loki-index-gateway/0.log" Dec 08 20:33:22 crc kubenswrapper[4706]: I1208 20:33:22.083471 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_9a912025-8f8e-4d61-861e-c933dafe4660/loki-ingester/0.log" Dec 08 20:33:22 crc kubenswrapper[4706]: I1208 20:33:22.224134 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-7c8cd744d9-xwfnt_fbd5715c-12b1-4660-bdd6-39087b0ae32a/loki-query-frontend/0.log" Dec 08 20:33:22 crc kubenswrapper[4706]: I1208 20:33:22.708149 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-5467947bf7-dxbfm_a9df4fe8-4ed5-4bd9-a514-9ed4259283b1/loki-querier/0.log" Dec 08 20:33:22 crc kubenswrapper[4706]: I1208 20:33:22.874768 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-sh8fx_784135f7-ff54-43df-b7e7-ea4f0b940fdc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:23 crc kubenswrapper[4706]: I1208 20:33:23.205496 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-swwvh_9576d118-0baa-462a-b75a-a31af5b84c66/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:23 crc kubenswrapper[4706]: I1208 20:33:23.262299 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-862hf_797f5d64-d90a-4466-95f7-a9b4b43d9f1c/init/0.log" Dec 08 20:33:23 crc kubenswrapper[4706]: I1208 20:33:23.488534 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-862hf_797f5d64-d90a-4466-95f7-a9b4b43d9f1c/init/0.log" Dec 08 20:33:23 crc kubenswrapper[4706]: I1208 20:33:23.555524 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-kkq6s_a9260164-eeea-47a2-9502-1744cba85586/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:23 crc kubenswrapper[4706]: I1208 20:33:23.671952 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-862hf_797f5d64-d90a-4466-95f7-a9b4b43d9f1c/dnsmasq-dns/0.log" Dec 08 20:33:24 crc kubenswrapper[4706]: I1208 20:33:24.397777 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8424f86e-b640-44f6-abd9-dc8c7a5a55cc/glance-httpd/0.log" Dec 08 20:33:24 crc kubenswrapper[4706]: I1208 20:33:24.451542 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8424f86e-b640-44f6-abd9-dc8c7a5a55cc/glance-log/0.log" Dec 08 20:33:24 crc kubenswrapper[4706]: I1208 20:33:24.640887 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6249f0fa-1385-4546-89c3-60a91e24da43/glance-httpd/0.log" Dec 08 20:33:24 crc kubenswrapper[4706]: I1208 20:33:24.738149 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6249f0fa-1385-4546-89c3-60a91e24da43/glance-log/0.log" Dec 08 20:33:24 crc kubenswrapper[4706]: I1208 20:33:24.787570 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-g4kg4_42663da5-6fa0-4332-8222-63fd50d7484e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:25 crc kubenswrapper[4706]: I1208 20:33:25.098015 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-zjn26_e30a59fd-9346-4fae-8a9d-14c8700fd608/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:25 crc kubenswrapper[4706]: I1208 20:33:25.406882 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29420401-b95cm_68bc049a-f060-4de4-8be5-c88c55abebd9/keystone-cron/0.log" Dec 08 20:33:25 crc kubenswrapper[4706]: I1208 20:33:25.543629 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-786cfc44fd-zdf7t_93159ff6-eda3-42c1-94db-46a5d0777a47/keystone-api/0.log" Dec 08 20:33:25 crc kubenswrapper[4706]: I1208 20:33:25.666526 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_aa4c62bf-4f94-4e1e-afd5-9036da150cc4/kube-state-metrics/0.log" Dec 08 20:33:25 crc kubenswrapper[4706]: I1208 20:33:25.976387 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-sw9tn_93b3bae1-83c6-40b3-8c4a-e5d4fc586bf8/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:27 crc kubenswrapper[4706]: I1208 20:33:27.113290 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79bc9d977-gw6pk_94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43/neutron-httpd/0.log" Dec 08 20:33:27 crc kubenswrapper[4706]: I1208 20:33:27.220403 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79bc9d977-gw6pk_94b7ac1c-e1fb-4d04-8881-4b3d1d7b6e43/neutron-api/0.log" Dec 08 20:33:27 crc kubenswrapper[4706]: I1208 20:33:27.380884 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9lfkb_543c86ef-f2f1-4b75-9de2-d613819b3e9a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:28 crc kubenswrapper[4706]: I1208 20:33:28.144637 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_48240f0b-4bff-4e59-b151-7e2922ee95f0/nova-api-log/0.log" Dec 08 20:33:28 crc kubenswrapper[4706]: I1208 20:33:28.427957 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_667cc0c8-9292-4cbd-a013-bd645c4b57d7/nova-cell0-conductor-conductor/0.log" Dec 08 20:33:28 crc kubenswrapper[4706]: I1208 20:33:28.750249 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_48240f0b-4bff-4e59-b151-7e2922ee95f0/nova-api-api/0.log" Dec 08 20:33:28 crc kubenswrapper[4706]: I1208 20:33:28.966979 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_734cc0d6-b154-4980-afbe-67d4d1b9b675/nova-cell1-conductor-conductor/0.log" Dec 08 20:33:29 crc kubenswrapper[4706]: I1208 20:33:29.153011 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8d07dfb3-a2ed-4e66-80a3-3288ff3f8669/nova-cell1-novncproxy-novncproxy/0.log" Dec 08 20:33:29 crc kubenswrapper[4706]: I1208 20:33:29.286311 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-m2jvl_43b41201-b854-4025-b70b-4455fcd8fc3c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:29 crc kubenswrapper[4706]: I1208 20:33:29.615491 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7eb161ee-d8ee-45e4-9fce-a0661293603a/nova-metadata-log/0.log" Dec 08 20:33:30 crc kubenswrapper[4706]: I1208 20:33:30.239597 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a0ae432b-2caa-458f-9ac4-5172e21ef898/nova-scheduler-scheduler/0.log" Dec 08 20:33:30 crc kubenswrapper[4706]: I1208 20:33:30.371893 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_13532f07-b125-4d67-bdb8-006b583bea40/mysql-bootstrap/0.log" Dec 08 20:33:30 crc kubenswrapper[4706]: I1208 20:33:30.484745 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_ad4728d1-aa2e-4ab5-81dc-22f56a79e874/cloudkitty-proc/0.log" Dec 08 20:33:30 crc kubenswrapper[4706]: I1208 20:33:30.699178 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_13532f07-b125-4d67-bdb8-006b583bea40/mysql-bootstrap/0.log" Dec 08 20:33:30 crc kubenswrapper[4706]: I1208 20:33:30.706841 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_13532f07-b125-4d67-bdb8-006b583bea40/galera/0.log" Dec 08 20:33:31 crc kubenswrapper[4706]: I1208 20:33:31.037895 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_156710b7-7dbe-4bb7-9c69-ce726d4b7a4c/mysql-bootstrap/0.log" Dec 08 20:33:31 crc kubenswrapper[4706]: I1208 20:33:31.223025 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_156710b7-7dbe-4bb7-9c69-ce726d4b7a4c/mysql-bootstrap/0.log" Dec 08 20:33:31 crc kubenswrapper[4706]: I1208 20:33:31.305038 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_156710b7-7dbe-4bb7-9c69-ce726d4b7a4c/galera/0.log" Dec 08 20:33:31 crc kubenswrapper[4706]: I1208 20:33:31.514937 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7eb161ee-d8ee-45e4-9fce-a0661293603a/nova-metadata-metadata/0.log" Dec 08 20:33:31 crc kubenswrapper[4706]: I1208 20:33:31.558682 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1b714f3f-e93c-4a5b-ae82-1f551b163fe6/openstackclient/0.log" Dec 08 20:33:31 crc kubenswrapper[4706]: I1208 20:33:31.688787 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-78wps_79f551cd-d08d-4739-9041-4543f05fbb60/ovn-controller/0.log" Dec 08 20:33:31 crc kubenswrapper[4706]: I1208 20:33:31.786568 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-schr5_a00cac1b-92f7-49a1-bbdc-97c15777b09a/openstack-network-exporter/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.006591 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb6gx_25287a9f-27c5-47a7-945e-684eea0157c8/ovsdb-server-init/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.266925 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb6gx_25287a9f-27c5-47a7-945e-684eea0157c8/ovsdb-server-init/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.332735 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb6gx_25287a9f-27c5-47a7-945e-684eea0157c8/ovs-vswitchd/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.342766 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb6gx_25287a9f-27c5-47a7-945e-684eea0157c8/ovsdb-server/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.543407 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-zxsvx_2c0f5bdb-fa5e-41b2-b271-f6e6d3f94e78/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.596125 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_240f9d54-934b-4d65-976e-a0a5a9ca7908/openstack-network-exporter/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.746602 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_240f9d54-934b-4d65-976e-a0a5a9ca7908/ovn-northd/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.889834 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f498bda7-c379-4808-bd6c-8e5513c4322d/openstack-network-exporter/0.log" Dec 08 20:33:32 crc kubenswrapper[4706]: I1208 20:33:32.964172 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f498bda7-c379-4808-bd6c-8e5513c4322d/ovsdbserver-nb/0.log" Dec 08 20:33:33 crc kubenswrapper[4706]: I1208 20:33:33.182808 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_40f5d973-ac12-469a-a1d9-aa871b7cc9a3/openstack-network-exporter/0.log" Dec 08 20:33:33 crc kubenswrapper[4706]: I1208 20:33:33.199795 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_40f5d973-ac12-469a-a1d9-aa871b7cc9a3/ovsdbserver-sb/0.log" Dec 08 20:33:33 crc kubenswrapper[4706]: I1208 20:33:33.806412 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7494d94d4d-hn8rd_4abc8b62-e549-41c9-8bc9-8d35f3074e79/placement-api/0.log" Dec 08 20:33:33 crc kubenswrapper[4706]: I1208 20:33:33.877382 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7494d94d4d-hn8rd_4abc8b62-e549-41c9-8bc9-8d35f3074e79/placement-log/0.log" Dec 08 20:33:33 crc kubenswrapper[4706]: I1208 20:33:33.973215 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/init-config-reloader/0.log" Dec 08 20:33:34 crc kubenswrapper[4706]: I1208 20:33:34.238711 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/init-config-reloader/0.log" Dec 08 20:33:34 crc kubenswrapper[4706]: I1208 20:33:34.329959 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/config-reloader/0.log" Dec 08 20:33:34 crc kubenswrapper[4706]: I1208 20:33:34.357238 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/thanos-sidecar/0.log" Dec 08 20:33:34 crc kubenswrapper[4706]: I1208 20:33:34.366499 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_eeb4b596-a4eb-4e91-bee9-123f57198cd3/prometheus/0.log" Dec 08 20:33:34 crc kubenswrapper[4706]: I1208 20:33:34.609445 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ecc9ad61-4d9c-4d26-a240-398cfc8c3165/setup-container/0.log" Dec 08 20:33:34 crc kubenswrapper[4706]: I1208 20:33:34.896140 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ecc9ad61-4d9c-4d26-a240-398cfc8c3165/setup-container/0.log" Dec 08 20:33:34 crc kubenswrapper[4706]: I1208 20:33:34.923080 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_faed709c-224b-4c30-9af3-d64d34525599/setup-container/0.log" Dec 08 20:33:34 crc kubenswrapper[4706]: I1208 20:33:34.937194 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ecc9ad61-4d9c-4d26-a240-398cfc8c3165/rabbitmq/0.log" Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.183208 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_faed709c-224b-4c30-9af3-d64d34525599/rabbitmq/0.log" Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.247558 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_faed709c-224b-4c30-9af3-d64d34525599/setup-container/0.log" Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.368779 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-sp25f_5d9ea4d5-f428-449d-baf9-880c8a2637aa/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.684996 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-44ss6_ad5354c5-7ee5-4e8a-8e20-84ac36e97e75/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.692390 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-hvs9v_074561a7-7190-4458-a9aa-1dac832175fb/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.839077 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.839146 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.900367 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jkcr5_893e5057-db47-44b6-b3da-f4e59a5992c5/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:35 crc kubenswrapper[4706]: I1208 20:33:35.976760 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-rhjj4_adf18455-0f5d-4d8e-a46d-c77523de9f84/ssh-known-hosts-edpm-deployment/0.log" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.210666 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lpnb8"] Dec 08 20:33:36 crc kubenswrapper[4706]: E1208 20:33:36.214712 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4206f9d-5fae-4379-9efa-b5380843d090" containerName="container-00" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.214750 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4206f9d-5fae-4379-9efa-b5380843d090" containerName="container-00" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.215102 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4206f9d-5fae-4379-9efa-b5380843d090" containerName="container-00" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.218746 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.260164 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lpnb8"] Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.326777 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzg9w\" (UniqueName: \"kubernetes.io/projected/b37d3642-1357-4be7-8d17-1d0db472f060-kube-api-access-dzg9w\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.327040 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-utilities\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.327177 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-catalog-content\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.436520 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d86ddd67f-5b6vx_637d7d72-7ba8-4a35-b37c-35aef7813c75/proxy-server/0.log" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.442851 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-utilities\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.443006 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-catalog-content\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.443067 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzg9w\" (UniqueName: \"kubernetes.io/projected/b37d3642-1357-4be7-8d17-1d0db472f060-kube-api-access-dzg9w\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.445373 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-utilities\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.445452 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-catalog-content\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.553492 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d86ddd67f-5b6vx_637d7d72-7ba8-4a35-b37c-35aef7813c75/proxy-httpd/0.log" Dec 08 20:33:36 crc kubenswrapper[4706]: I1208 20:33:36.659336 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-thlcc_a7d7a844-b2ae-477f-b8c0-d14f8ec87d3c/swift-ring-rebalance/0.log" Dec 08 20:33:37 crc kubenswrapper[4706]: I1208 20:33:37.250765 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzg9w\" (UniqueName: \"kubernetes.io/projected/b37d3642-1357-4be7-8d17-1d0db472f060-kube-api-access-dzg9w\") pod \"redhat-operators-lpnb8\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:37 crc kubenswrapper[4706]: I1208 20:33:37.498736 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/account-reaper/0.log" Dec 08 20:33:37 crc kubenswrapper[4706]: I1208 20:33:37.502717 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:37 crc kubenswrapper[4706]: I1208 20:33:37.582613 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/account-auditor/0.log" Dec 08 20:33:37 crc kubenswrapper[4706]: I1208 20:33:37.618965 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/account-replicator/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.010381 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/container-replicator/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.010756 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/account-server/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.127598 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/container-server/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.165518 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/container-auditor/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.181568 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lpnb8"] Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.378228 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/container-updater/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.420567 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-expirer/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.505924 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-auditor/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.582637 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-replicator/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.637779 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-server/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.665644 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/object-updater/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.767812 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/rsync/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.846830 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d81adfd7-689a-496a-b4bf-c35450b63ca9/swift-recon-cron/0.log" Dec 08 20:33:38 crc kubenswrapper[4706]: I1208 20:33:38.963583 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-f6l4p_39dc4f79-e40f-479c-b82b-8ac18056c9ec/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:39 crc kubenswrapper[4706]: I1208 20:33:39.042194 4706 generic.go:334] "Generic (PLEG): container finished" podID="b37d3642-1357-4be7-8d17-1d0db472f060" containerID="9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643" exitCode=0 Dec 08 20:33:39 crc kubenswrapper[4706]: I1208 20:33:39.042287 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpnb8" event={"ID":"b37d3642-1357-4be7-8d17-1d0db472f060","Type":"ContainerDied","Data":"9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643"} Dec 08 20:33:39 crc kubenswrapper[4706]: I1208 20:33:39.042334 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpnb8" event={"ID":"b37d3642-1357-4be7-8d17-1d0db472f060","Type":"ContainerStarted","Data":"d98159869207c419d97a36217b480652951403e14eade09a96a42a7baa3182d9"} Dec 08 20:33:39 crc kubenswrapper[4706]: I1208 20:33:39.206162 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_e673aa84-089b-4ab2-abcd-9c3827d97cbf/tempest-tests-tempest-tests-runner/0.log" Dec 08 20:33:39 crc kubenswrapper[4706]: I1208 20:33:39.872509 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_4fbf8282-ee38-4ca3-a1e5-fa89ffd85f4a/test-operator-logs-container/0.log" Dec 08 20:33:39 crc kubenswrapper[4706]: I1208 20:33:39.969940 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-x6f2s_7ea46b46-9aa7-46f3-a235-8eb100586d2f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.573770 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5rd8s"] Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.588981 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.590102 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rd8s"] Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.654035 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w4zb\" (UniqueName: \"kubernetes.io/projected/495b7e1d-73b6-43d8-a009-314c3985d44e-kube-api-access-9w4zb\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.654202 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-utilities\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.654385 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-catalog-content\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.756963 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w4zb\" (UniqueName: \"kubernetes.io/projected/495b7e1d-73b6-43d8-a009-314c3985d44e-kube-api-access-9w4zb\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.757050 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-utilities\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.757229 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-catalog-content\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.757716 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-catalog-content\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.758298 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-utilities\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.785142 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w4zb\" (UniqueName: \"kubernetes.io/projected/495b7e1d-73b6-43d8-a009-314c3985d44e-kube-api-access-9w4zb\") pod \"community-operators-5rd8s\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:40 crc kubenswrapper[4706]: I1208 20:33:40.918034 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:41 crc kubenswrapper[4706]: I1208 20:33:41.106155 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpnb8" event={"ID":"b37d3642-1357-4be7-8d17-1d0db472f060","Type":"ContainerStarted","Data":"41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5"} Dec 08 20:33:41 crc kubenswrapper[4706]: I1208 20:33:41.691374 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rd8s"] Dec 08 20:33:42 crc kubenswrapper[4706]: I1208 20:33:42.118677 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rd8s" event={"ID":"495b7e1d-73b6-43d8-a009-314c3985d44e","Type":"ContainerStarted","Data":"f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83"} Dec 08 20:33:42 crc kubenswrapper[4706]: I1208 20:33:42.119293 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rd8s" event={"ID":"495b7e1d-73b6-43d8-a009-314c3985d44e","Type":"ContainerStarted","Data":"3f6dd1d6905817964d0c43f5057745c5715ae365e2b4442edd4f3320987853b1"} Dec 08 20:33:43 crc kubenswrapper[4706]: I1208 20:33:43.134962 4706 generic.go:334] "Generic (PLEG): container finished" podID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerID="f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83" exitCode=0 Dec 08 20:33:43 crc kubenswrapper[4706]: I1208 20:33:43.135012 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rd8s" event={"ID":"495b7e1d-73b6-43d8-a009-314c3985d44e","Type":"ContainerDied","Data":"f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83"} Dec 08 20:33:43 crc kubenswrapper[4706]: I1208 20:33:43.648519 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fd927fec-2b54-41d2-affb-19f9fd8c96ab/memcached/0.log" Dec 08 20:33:44 crc kubenswrapper[4706]: I1208 20:33:44.154642 4706 generic.go:334] "Generic (PLEG): container finished" podID="b37d3642-1357-4be7-8d17-1d0db472f060" containerID="41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5" exitCode=0 Dec 08 20:33:44 crc kubenswrapper[4706]: I1208 20:33:44.154808 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpnb8" event={"ID":"b37d3642-1357-4be7-8d17-1d0db472f060","Type":"ContainerDied","Data":"41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5"} Dec 08 20:33:45 crc kubenswrapper[4706]: I1208 20:33:45.172740 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpnb8" event={"ID":"b37d3642-1357-4be7-8d17-1d0db472f060","Type":"ContainerStarted","Data":"6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1"} Dec 08 20:33:45 crc kubenswrapper[4706]: I1208 20:33:45.178349 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rd8s" event={"ID":"495b7e1d-73b6-43d8-a009-314c3985d44e","Type":"ContainerStarted","Data":"a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0"} Dec 08 20:33:45 crc kubenswrapper[4706]: I1208 20:33:45.200957 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lpnb8" podStartSLOduration=3.607824408 podStartE2EDuration="9.200937217s" podCreationTimestamp="2025-12-08 20:33:36 +0000 UTC" firstStartedPulling="2025-12-08 20:33:39.045404513 +0000 UTC m=+4301.687605516" lastFinishedPulling="2025-12-08 20:33:44.638517322 +0000 UTC m=+4307.280718325" observedRunningTime="2025-12-08 20:33:45.198413205 +0000 UTC m=+4307.840614218" watchObservedRunningTime="2025-12-08 20:33:45.200937217 +0000 UTC m=+4307.843138220" Dec 08 20:33:46 crc kubenswrapper[4706]: I1208 20:33:46.199770 4706 generic.go:334] "Generic (PLEG): container finished" podID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerID="a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0" exitCode=0 Dec 08 20:33:46 crc kubenswrapper[4706]: I1208 20:33:46.199913 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rd8s" event={"ID":"495b7e1d-73b6-43d8-a009-314c3985d44e","Type":"ContainerDied","Data":"a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0"} Dec 08 20:33:47 crc kubenswrapper[4706]: I1208 20:33:47.213621 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rd8s" event={"ID":"495b7e1d-73b6-43d8-a009-314c3985d44e","Type":"ContainerStarted","Data":"85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f"} Dec 08 20:33:47 crc kubenswrapper[4706]: I1208 20:33:47.239811 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5rd8s" podStartSLOduration=3.7567569709999997 podStartE2EDuration="7.239789437s" podCreationTimestamp="2025-12-08 20:33:40 +0000 UTC" firstStartedPulling="2025-12-08 20:33:43.137893175 +0000 UTC m=+4305.780094178" lastFinishedPulling="2025-12-08 20:33:46.620925641 +0000 UTC m=+4309.263126644" observedRunningTime="2025-12-08 20:33:47.232717065 +0000 UTC m=+4309.874918078" watchObservedRunningTime="2025-12-08 20:33:47.239789437 +0000 UTC m=+4309.881990440" Dec 08 20:33:47 crc kubenswrapper[4706]: I1208 20:33:47.504054 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:47 crc kubenswrapper[4706]: I1208 20:33:47.504112 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:33:48 crc kubenswrapper[4706]: I1208 20:33:48.562067 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lpnb8" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="registry-server" probeResult="failure" output=< Dec 08 20:33:48 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 20:33:48 crc kubenswrapper[4706]: > Dec 08 20:33:50 crc kubenswrapper[4706]: I1208 20:33:50.919697 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:50 crc kubenswrapper[4706]: I1208 20:33:50.920357 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:50 crc kubenswrapper[4706]: I1208 20:33:50.974505 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:51 crc kubenswrapper[4706]: I1208 20:33:51.339572 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:51 crc kubenswrapper[4706]: I1208 20:33:51.744804 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rd8s"] Dec 08 20:33:53 crc kubenswrapper[4706]: I1208 20:33:53.286575 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5rd8s" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerName="registry-server" containerID="cri-o://85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f" gracePeriod=2 Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.119757 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.250363 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-catalog-content\") pod \"495b7e1d-73b6-43d8-a009-314c3985d44e\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.250438 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w4zb\" (UniqueName: \"kubernetes.io/projected/495b7e1d-73b6-43d8-a009-314c3985d44e-kube-api-access-9w4zb\") pod \"495b7e1d-73b6-43d8-a009-314c3985d44e\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.250477 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-utilities\") pod \"495b7e1d-73b6-43d8-a009-314c3985d44e\" (UID: \"495b7e1d-73b6-43d8-a009-314c3985d44e\") " Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.251469 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-utilities" (OuterVolumeSpecName: "utilities") pod "495b7e1d-73b6-43d8-a009-314c3985d44e" (UID: "495b7e1d-73b6-43d8-a009-314c3985d44e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.259670 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/495b7e1d-73b6-43d8-a009-314c3985d44e-kube-api-access-9w4zb" (OuterVolumeSpecName: "kube-api-access-9w4zb") pod "495b7e1d-73b6-43d8-a009-314c3985d44e" (UID: "495b7e1d-73b6-43d8-a009-314c3985d44e"). InnerVolumeSpecName "kube-api-access-9w4zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.322187 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "495b7e1d-73b6-43d8-a009-314c3985d44e" (UID: "495b7e1d-73b6-43d8-a009-314c3985d44e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.324111 4706 generic.go:334] "Generic (PLEG): container finished" podID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerID="85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f" exitCode=0 Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.324163 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rd8s" event={"ID":"495b7e1d-73b6-43d8-a009-314c3985d44e","Type":"ContainerDied","Data":"85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f"} Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.324203 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rd8s" event={"ID":"495b7e1d-73b6-43d8-a009-314c3985d44e","Type":"ContainerDied","Data":"3f6dd1d6905817964d0c43f5057745c5715ae365e2b4442edd4f3320987853b1"} Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.324223 4706 scope.go:117] "RemoveContainer" containerID="85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.324217 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rd8s" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.396972 4706 scope.go:117] "RemoveContainer" containerID="a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.399848 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.399879 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w4zb\" (UniqueName: \"kubernetes.io/projected/495b7e1d-73b6-43d8-a009-314c3985d44e-kube-api-access-9w4zb\") on node \"crc\" DevicePath \"\"" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.399891 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495b7e1d-73b6-43d8-a009-314c3985d44e-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.453678 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rd8s"] Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.478506 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5rd8s"] Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.488421 4706 scope.go:117] "RemoveContainer" containerID="f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.540478 4706 scope.go:117] "RemoveContainer" containerID="85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f" Dec 08 20:33:54 crc kubenswrapper[4706]: E1208 20:33:54.541739 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f\": container with ID starting with 85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f not found: ID does not exist" containerID="85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.541811 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f"} err="failed to get container status \"85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f\": rpc error: code = NotFound desc = could not find container \"85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f\": container with ID starting with 85bca9cca7b4111954cd036a70acda8de7ab73c806efcb8273593d4c4151164f not found: ID does not exist" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.541884 4706 scope.go:117] "RemoveContainer" containerID="a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0" Dec 08 20:33:54 crc kubenswrapper[4706]: E1208 20:33:54.542856 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0\": container with ID starting with a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0 not found: ID does not exist" containerID="a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.542921 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0"} err="failed to get container status \"a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0\": rpc error: code = NotFound desc = could not find container \"a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0\": container with ID starting with a1bbf0859c4efa5e03a029747f724bea784d1e63c61ad8e193a98e8cdc5a30d0 not found: ID does not exist" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.542958 4706 scope.go:117] "RemoveContainer" containerID="f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83" Dec 08 20:33:54 crc kubenswrapper[4706]: E1208 20:33:54.543308 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83\": container with ID starting with f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83 not found: ID does not exist" containerID="f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83" Dec 08 20:33:54 crc kubenswrapper[4706]: I1208 20:33:54.543331 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83"} err="failed to get container status \"f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83\": rpc error: code = NotFound desc = could not find container \"f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83\": container with ID starting with f019a848c1c02c6087426524c703de26d7f1fdbdd7277059875ec3a5bf5fda83 not found: ID does not exist" Dec 08 20:33:55 crc kubenswrapper[4706]: I1208 20:33:55.621826 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" path="/var/lib/kubelet/pods/495b7e1d-73b6-43d8-a009-314c3985d44e/volumes" Dec 08 20:33:58 crc kubenswrapper[4706]: I1208 20:33:58.561747 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lpnb8" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="registry-server" probeResult="failure" output=< Dec 08 20:33:58 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 08 20:33:58 crc kubenswrapper[4706]: > Dec 08 20:34:05 crc kubenswrapper[4706]: I1208 20:34:05.836517 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:34:05 crc kubenswrapper[4706]: I1208 20:34:05.837435 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:34:07 crc kubenswrapper[4706]: I1208 20:34:07.563209 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:34:07 crc kubenswrapper[4706]: I1208 20:34:07.637814 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:34:07 crc kubenswrapper[4706]: I1208 20:34:07.811746 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lpnb8"] Dec 08 20:34:09 crc kubenswrapper[4706]: I1208 20:34:09.490919 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lpnb8" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="registry-server" containerID="cri-o://6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1" gracePeriod=2 Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.149715 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.225813 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzg9w\" (UniqueName: \"kubernetes.io/projected/b37d3642-1357-4be7-8d17-1d0db472f060-kube-api-access-dzg9w\") pod \"b37d3642-1357-4be7-8d17-1d0db472f060\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.226094 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-catalog-content\") pod \"b37d3642-1357-4be7-8d17-1d0db472f060\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.226342 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-utilities\") pod \"b37d3642-1357-4be7-8d17-1d0db472f060\" (UID: \"b37d3642-1357-4be7-8d17-1d0db472f060\") " Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.227050 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-utilities" (OuterVolumeSpecName: "utilities") pod "b37d3642-1357-4be7-8d17-1d0db472f060" (UID: "b37d3642-1357-4be7-8d17-1d0db472f060"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.233596 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b37d3642-1357-4be7-8d17-1d0db472f060-kube-api-access-dzg9w" (OuterVolumeSpecName: "kube-api-access-dzg9w") pod "b37d3642-1357-4be7-8d17-1d0db472f060" (UID: "b37d3642-1357-4be7-8d17-1d0db472f060"). InnerVolumeSpecName "kube-api-access-dzg9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.330373 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.330437 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzg9w\" (UniqueName: \"kubernetes.io/projected/b37d3642-1357-4be7-8d17-1d0db472f060-kube-api-access-dzg9w\") on node \"crc\" DevicePath \"\"" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.352620 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b37d3642-1357-4be7-8d17-1d0db472f060" (UID: "b37d3642-1357-4be7-8d17-1d0db472f060"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.433137 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b37d3642-1357-4be7-8d17-1d0db472f060-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.506590 4706 generic.go:334] "Generic (PLEG): container finished" podID="b37d3642-1357-4be7-8d17-1d0db472f060" containerID="6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1" exitCode=0 Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.506646 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpnb8" event={"ID":"b37d3642-1357-4be7-8d17-1d0db472f060","Type":"ContainerDied","Data":"6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1"} Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.506681 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpnb8" event={"ID":"b37d3642-1357-4be7-8d17-1d0db472f060","Type":"ContainerDied","Data":"d98159869207c419d97a36217b480652951403e14eade09a96a42a7baa3182d9"} Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.506706 4706 scope.go:117] "RemoveContainer" containerID="6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.506721 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpnb8" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.550549 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lpnb8"] Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.567152 4706 scope.go:117] "RemoveContainer" containerID="41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.571954 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lpnb8"] Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.612889 4706 scope.go:117] "RemoveContainer" containerID="9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.671364 4706 scope.go:117] "RemoveContainer" containerID="6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1" Dec 08 20:34:10 crc kubenswrapper[4706]: E1208 20:34:10.672056 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1\": container with ID starting with 6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1 not found: ID does not exist" containerID="6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.672105 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1"} err="failed to get container status \"6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1\": rpc error: code = NotFound desc = could not find container \"6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1\": container with ID starting with 6a0f95436c03f124f90278b8b5ca2c95ac868036a757220185c3ce45ebe76fd1 not found: ID does not exist" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.672149 4706 scope.go:117] "RemoveContainer" containerID="41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5" Dec 08 20:34:10 crc kubenswrapper[4706]: E1208 20:34:10.672638 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5\": container with ID starting with 41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5 not found: ID does not exist" containerID="41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.672666 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5"} err="failed to get container status \"41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5\": rpc error: code = NotFound desc = could not find container \"41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5\": container with ID starting with 41b8903e9f7d687469ca531f045bc8bbcaa300f15735bedc72a63070a0ed5cb5 not found: ID does not exist" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.672683 4706 scope.go:117] "RemoveContainer" containerID="9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643" Dec 08 20:34:10 crc kubenswrapper[4706]: E1208 20:34:10.673235 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643\": container with ID starting with 9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643 not found: ID does not exist" containerID="9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643" Dec 08 20:34:10 crc kubenswrapper[4706]: I1208 20:34:10.673331 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643"} err="failed to get container status \"9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643\": rpc error: code = NotFound desc = could not find container \"9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643\": container with ID starting with 9d58e4fd06829aec88a238336d1a251e7cf9a430745faeaa25272f8397a3f643 not found: ID does not exist" Dec 08 20:34:11 crc kubenswrapper[4706]: I1208 20:34:11.625759 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" path="/var/lib/kubelet/pods/b37d3642-1357-4be7-8d17-1d0db472f060/volumes" Dec 08 20:34:14 crc kubenswrapper[4706]: I1208 20:34:14.182308 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-qtnlw_4c42f8b9-a85c-4a45-a8f8-4a43e561a14e/manager/0.log" Dec 08 20:34:14 crc kubenswrapper[4706]: I1208 20:34:14.208633 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-qtnlw_4c42f8b9-a85c-4a45-a8f8-4a43e561a14e/kube-rbac-proxy/0.log" Dec 08 20:34:14 crc kubenswrapper[4706]: I1208 20:34:14.518372 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-j2mjl_e5899e67-6618-43f5-8454-fddfae494e4a/manager/0.log" Dec 08 20:34:14 crc kubenswrapper[4706]: I1208 20:34:14.625174 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-j2mjl_e5899e67-6618-43f5-8454-fddfae494e4a/kube-rbac-proxy/0.log" Dec 08 20:34:14 crc kubenswrapper[4706]: I1208 20:34:14.736457 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/util/0.log" Dec 08 20:34:15 crc kubenswrapper[4706]: I1208 20:34:15.561055 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/util/0.log" Dec 08 20:34:15 crc kubenswrapper[4706]: I1208 20:34:15.592826 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/pull/0.log" Dec 08 20:34:15 crc kubenswrapper[4706]: I1208 20:34:15.597570 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/pull/0.log" Dec 08 20:34:15 crc kubenswrapper[4706]: I1208 20:34:15.790552 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/util/0.log" Dec 08 20:34:15 crc kubenswrapper[4706]: I1208 20:34:15.791232 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/pull/0.log" Dec 08 20:34:15 crc kubenswrapper[4706]: I1208 20:34:15.830016 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dae59db2bf4c02e33f4395cd1388e2543c0987be5a06066cfa81f23fad9gdt8_3231e588-6c34-4c26-9909-ece08e2661c3/extract/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.048641 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-7fvjs_0f144ce9-7cc0-41f3-84ea-654579d32c0b/kube-rbac-proxy/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.074237 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-7fvjs_0f144ce9-7cc0-41f3-84ea-654579d32c0b/manager/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.205275 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-b7m9l_b02b307a-7d4c-4d20-b388-18f6fc6ec0ee/kube-rbac-proxy/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.378075 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-b7m9l_b02b307a-7d4c-4d20-b388-18f6fc6ec0ee/manager/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.442424 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-cmg9l_2acd6f54-87a9-4e50-859c-42a7c8e9d098/kube-rbac-proxy/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.501281 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-cmg9l_2acd6f54-87a9-4e50-859c-42a7c8e9d098/manager/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.654667 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cggng_1290df78-3e2c-494e-a264-c495d80c0bef/kube-rbac-proxy/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.706172 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cggng_1290df78-3e2c-494e-a264-c495d80c0bef/manager/0.log" Dec 08 20:34:16 crc kubenswrapper[4706]: I1208 20:34:16.956975 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-qmcdh_ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3/kube-rbac-proxy/0.log" Dec 08 20:34:17 crc kubenswrapper[4706]: I1208 20:34:17.112133 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5cwxk_42d526f1-2349-4aef-80fa-a6e59f3630e9/kube-rbac-proxy/0.log" Dec 08 20:34:17 crc kubenswrapper[4706]: I1208 20:34:17.150594 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-qmcdh_ca9ffe3e-a317-4570-8a00-a30ac2dfb1f3/manager/0.log" Dec 08 20:34:17 crc kubenswrapper[4706]: I1208 20:34:17.505140 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5cwxk_42d526f1-2349-4aef-80fa-a6e59f3630e9/manager/0.log" Dec 08 20:34:17 crc kubenswrapper[4706]: I1208 20:34:17.797652 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kg2wq_fab26e20-8c1c-4a9e-b7c4-8d4815a1568e/kube-rbac-proxy/0.log" Dec 08 20:34:17 crc kubenswrapper[4706]: I1208 20:34:17.872008 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-pfq9b_53d61380-3a1e-45e5-a039-9d339bcdb9ea/kube-rbac-proxy/0.log" Dec 08 20:34:17 crc kubenswrapper[4706]: I1208 20:34:17.883403 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kg2wq_fab26e20-8c1c-4a9e-b7c4-8d4815a1568e/manager/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.019413 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-pfq9b_53d61380-3a1e-45e5-a039-9d339bcdb9ea/manager/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.100145 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-l8qxc_8ddb889c-1432-4225-a110-e0aabc0ce3bc/kube-rbac-proxy/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.145996 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-l8qxc_8ddb889c-1432-4225-a110-e0aabc0ce3bc/manager/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.318185 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-5xgcd_1f3c9d68-2093-45a5-886e-f19c43b3e8e9/kube-rbac-proxy/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.381554 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-58rkn_6da8aab3-b971-413d-8557-cdfad2c95d2c/kube-rbac-proxy/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.456970 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-5xgcd_1f3c9d68-2093-45a5-886e-f19c43b3e8e9/manager/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.633455 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-58rkn_6da8aab3-b971-413d-8557-cdfad2c95d2c/manager/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.737833 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5ndn8_eedc4c77-93bc-40b9-ba3c-e7222d8828b5/kube-rbac-proxy/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.792084 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5ndn8_eedc4c77-93bc-40b9-ba3c-e7222d8828b5/manager/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.928999 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f6nqtx_4fca7ad0-1e55-4b7a-bee5-08521a1836a2/manager/0.log" Dec 08 20:34:18 crc kubenswrapper[4706]: I1208 20:34:18.931598 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f6nqtx_4fca7ad0-1e55-4b7a-bee5-08521a1836a2/kube-rbac-proxy/0.log" Dec 08 20:34:19 crc kubenswrapper[4706]: I1208 20:34:19.261916 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-c5fr6_ca6c738d-aa2e-4be3-9827-fc03af84458d/registry-server/0.log" Dec 08 20:34:19 crc kubenswrapper[4706]: I1208 20:34:19.529062 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-vpwqp_ec6df814-0549-46a9-a9f3-2f94083bda52/kube-rbac-proxy/0.log" Dec 08 20:34:19 crc kubenswrapper[4706]: I1208 20:34:19.559357 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d96469866-85q7p_ea53cbe8-559b-45be-8ae8-25aeac554de4/operator/0.log" Dec 08 20:34:19 crc kubenswrapper[4706]: I1208 20:34:19.595861 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-vpwqp_ec6df814-0549-46a9-a9f3-2f94083bda52/manager/0.log" Dec 08 20:34:19 crc kubenswrapper[4706]: I1208 20:34:19.907753 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nbcmr_22335991-2ba5-4fba-a1d1-b391e4767961/kube-rbac-proxy/0.log" Dec 08 20:34:19 crc kubenswrapper[4706]: I1208 20:34:19.952356 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nbcmr_22335991-2ba5-4fba-a1d1-b391e4767961/manager/0.log" Dec 08 20:34:20 crc kubenswrapper[4706]: I1208 20:34:20.053752 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-bqsg2_9cca1f0f-9037-49d5-b8da-392421a67702/operator/0.log" Dec 08 20:34:20 crc kubenswrapper[4706]: I1208 20:34:20.229456 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-7htmr_e9c55dcb-b67a-43f9-9a80-8c38e7f94034/kube-rbac-proxy/0.log" Dec 08 20:34:20 crc kubenswrapper[4706]: I1208 20:34:20.290318 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-7htmr_e9c55dcb-b67a-43f9-9a80-8c38e7f94034/manager/0.log" Dec 08 20:34:20 crc kubenswrapper[4706]: I1208 20:34:20.487856 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5cbd74b459-5l5k7_beda813f-668e-45e8-8d1c-7c72caa6f294/kube-rbac-proxy/0.log" Dec 08 20:34:20 crc kubenswrapper[4706]: I1208 20:34:20.705902 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-qfsq9_375c721a-f040-42a8-9b43-f609258063a1/kube-rbac-proxy/0.log" Dec 08 20:34:20 crc kubenswrapper[4706]: I1208 20:34:20.792330 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-qfsq9_375c721a-f040-42a8-9b43-f609258063a1/manager/0.log" Dec 08 20:34:20 crc kubenswrapper[4706]: I1208 20:34:20.865335 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-b89665ccc-lvggm_61674e93-1252-4e28-ac66-b2a701745112/manager/0.log" Dec 08 20:34:21 crc kubenswrapper[4706]: I1208 20:34:21.053495 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-g6gls_ed8cb084-a66c-4fc0-afa1-068e58c80680/kube-rbac-proxy/0.log" Dec 08 20:34:21 crc kubenswrapper[4706]: I1208 20:34:21.138407 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5cbd74b459-5l5k7_beda813f-668e-45e8-8d1c-7c72caa6f294/manager/0.log" Dec 08 20:34:21 crc kubenswrapper[4706]: I1208 20:34:21.159096 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-g6gls_ed8cb084-a66c-4fc0-afa1-068e58c80680/manager/0.log" Dec 08 20:34:35 crc kubenswrapper[4706]: I1208 20:34:35.836351 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:34:35 crc kubenswrapper[4706]: I1208 20:34:35.837307 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:34:35 crc kubenswrapper[4706]: I1208 20:34:35.837388 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 20:34:35 crc kubenswrapper[4706]: I1208 20:34:35.838835 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4d17c26bf612a2b2ecf64b164dc8622c2374233ba25c61a4ad2d539b54439cf0"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 20:34:35 crc kubenswrapper[4706]: I1208 20:34:35.838916 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://4d17c26bf612a2b2ecf64b164dc8622c2374233ba25c61a4ad2d539b54439cf0" gracePeriod=600 Dec 08 20:34:36 crc kubenswrapper[4706]: I1208 20:34:36.844494 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="4d17c26bf612a2b2ecf64b164dc8622c2374233ba25c61a4ad2d539b54439cf0" exitCode=0 Dec 08 20:34:36 crc kubenswrapper[4706]: I1208 20:34:36.844609 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"4d17c26bf612a2b2ecf64b164dc8622c2374233ba25c61a4ad2d539b54439cf0"} Dec 08 20:34:36 crc kubenswrapper[4706]: I1208 20:34:36.845169 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerStarted","Data":"554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040"} Dec 08 20:34:36 crc kubenswrapper[4706]: I1208 20:34:36.845209 4706 scope.go:117] "RemoveContainer" containerID="67b7bc44b94071708565da6623ff77ba88330655b8246a4f0e68df3ed084f451" Dec 08 20:34:44 crc kubenswrapper[4706]: I1208 20:34:44.929203 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mxb2z_878ce0cf-8408-4bd3-90cc-948a66892c07/control-plane-machine-set-operator/0.log" Dec 08 20:34:45 crc kubenswrapper[4706]: I1208 20:34:45.137153 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j6s78_48960e60-2160-4afa-98ca-e67603fc9545/kube-rbac-proxy/0.log" Dec 08 20:34:45 crc kubenswrapper[4706]: I1208 20:34:45.196941 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j6s78_48960e60-2160-4afa-98ca-e67603fc9545/machine-api-operator/0.log" Dec 08 20:35:01 crc kubenswrapper[4706]: I1208 20:35:01.547950 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-c6s5g_8bf48929-3e5e-4fc7-a559-0ddcc2707cda/cert-manager-controller/0.log" Dec 08 20:35:01 crc kubenswrapper[4706]: I1208 20:35:01.692445 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-hzrrx_aefedb16-e712-45a1-8c7b-cd34002fd0bb/cert-manager-cainjector/0.log" Dec 08 20:35:01 crc kubenswrapper[4706]: I1208 20:35:01.781149 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-588tt_c1b92b7d-f282-496b-b69f-f55c0b9baec2/cert-manager-webhook/0.log" Dec 08 20:35:19 crc kubenswrapper[4706]: I1208 20:35:19.884905 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-224bh_0b6da606-8fe3-4228-81f7-533d8155c6f0/nmstate-console-plugin/0.log" Dec 08 20:35:20 crc kubenswrapper[4706]: I1208 20:35:20.167314 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-zqns4_1287503f-533b-4f53-b9b2-6792ad891790/nmstate-handler/0.log" Dec 08 20:35:20 crc kubenswrapper[4706]: I1208 20:35:20.280459 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-llfnx_05ea3b8c-286e-48e7-b232-ac550fcf165e/kube-rbac-proxy/0.log" Dec 08 20:35:20 crc kubenswrapper[4706]: I1208 20:35:20.305321 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-llfnx_05ea3b8c-286e-48e7-b232-ac550fcf165e/nmstate-metrics/0.log" Dec 08 20:35:20 crc kubenswrapper[4706]: I1208 20:35:20.695636 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-9jz6k_f56b4031-2db2-4d47-9299-9829b3963664/nmstate-operator/0.log" Dec 08 20:35:20 crc kubenswrapper[4706]: I1208 20:35:20.774925 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-kwnzx_aea9659b-8952-4f71-a8d5-970bfd9fffe0/nmstate-webhook/0.log" Dec 08 20:35:38 crc kubenswrapper[4706]: I1208 20:35:38.415526 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55474f77b6-v6n6q_b600c67d-b283-41ed-85f1-d19f3671f40f/kube-rbac-proxy/0.log" Dec 08 20:35:38 crc kubenswrapper[4706]: I1208 20:35:38.472615 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55474f77b6-v6n6q_b600c67d-b283-41ed-85f1-d19f3671f40f/manager/0.log" Dec 08 20:35:58 crc kubenswrapper[4706]: I1208 20:35:58.447814 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-9k84w_00d01f1b-f2d8-467d-94a3-e9a73966d334/kube-rbac-proxy/0.log" Dec 08 20:35:58 crc kubenswrapper[4706]: I1208 20:35:58.484198 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-9k84w_00d01f1b-f2d8-467d-94a3-e9a73966d334/controller/0.log" Dec 08 20:35:58 crc kubenswrapper[4706]: I1208 20:35:58.697871 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-frr-files/0.log" Dec 08 20:35:58 crc kubenswrapper[4706]: I1208 20:35:58.964883 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-reloader/0.log" Dec 08 20:35:58 crc kubenswrapper[4706]: I1208 20:35:58.985289 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-reloader/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.023869 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-frr-files/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.050735 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-metrics/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.323877 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-metrics/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.331678 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-frr-files/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.345176 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-reloader/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.688115 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-metrics/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.877051 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-frr-files/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.886449 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-reloader/0.log" Dec 08 20:35:59 crc kubenswrapper[4706]: I1208 20:35:59.898307 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/cp-metrics/0.log" Dec 08 20:36:00 crc kubenswrapper[4706]: I1208 20:36:00.003458 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/controller/0.log" Dec 08 20:36:00 crc kubenswrapper[4706]: I1208 20:36:00.096143 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/frr-metrics/0.log" Dec 08 20:36:00 crc kubenswrapper[4706]: I1208 20:36:00.214275 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/kube-rbac-proxy/0.log" Dec 08 20:36:00 crc kubenswrapper[4706]: I1208 20:36:00.319893 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/kube-rbac-proxy-frr/0.log" Dec 08 20:36:00 crc kubenswrapper[4706]: I1208 20:36:00.444518 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/reloader/0.log" Dec 08 20:36:00 crc kubenswrapper[4706]: I1208 20:36:00.683230 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-mdfmz_257f3480-2853-4319-b4c2-7eb65e6bb097/frr-k8s-webhook-server/0.log" Dec 08 20:36:00 crc kubenswrapper[4706]: I1208 20:36:00.994114 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-65c6ff64d8-bdjmj_fdc00423-9550-4ff3-84d9-78eeb50e29ea/manager/0.log" Dec 08 20:36:01 crc kubenswrapper[4706]: I1208 20:36:01.047862 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-75457555df-ll7rx_4ef64783-e4c8-4c02-884f-3dba766387b8/webhook-server/0.log" Dec 08 20:36:01 crc kubenswrapper[4706]: I1208 20:36:01.312310 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-49kgz_1398dd3e-9aac-4f48-b927-1862ed71c544/kube-rbac-proxy/0.log" Dec 08 20:36:01 crc kubenswrapper[4706]: I1208 20:36:01.940026 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jtzsv_1af879c2-1138-4ca6-92d5-013c50cb4ef2/frr/0.log" Dec 08 20:36:01 crc kubenswrapper[4706]: I1208 20:36:01.997310 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-49kgz_1398dd3e-9aac-4f48-b927-1862ed71c544/speaker/0.log" Dec 08 20:36:20 crc kubenswrapper[4706]: I1208 20:36:20.681250 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/util/0.log" Dec 08 20:36:20 crc kubenswrapper[4706]: I1208 20:36:20.856167 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/util/0.log" Dec 08 20:36:20 crc kubenswrapper[4706]: I1208 20:36:20.911395 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/pull/0.log" Dec 08 20:36:20 crc kubenswrapper[4706]: I1208 20:36:20.975656 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/pull/0.log" Dec 08 20:36:21 crc kubenswrapper[4706]: I1208 20:36:21.162177 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/pull/0.log" Dec 08 20:36:21 crc kubenswrapper[4706]: I1208 20:36:21.240724 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/util/0.log" Dec 08 20:36:21 crc kubenswrapper[4706]: I1208 20:36:21.255112 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgdmhv_03a7459f-3dd0-4cc8-9183-b82974d15a55/extract/0.log" Dec 08 20:36:21 crc kubenswrapper[4706]: I1208 20:36:21.457473 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/util/0.log" Dec 08 20:36:21 crc kubenswrapper[4706]: I1208 20:36:21.740019 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/pull/0.log" Dec 08 20:36:21 crc kubenswrapper[4706]: I1208 20:36:21.800087 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/pull/0.log" Dec 08 20:36:21 crc kubenswrapper[4706]: I1208 20:36:21.841884 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/util/0.log" Dec 08 20:36:22 crc kubenswrapper[4706]: I1208 20:36:22.385313 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/util/0.log" Dec 08 20:36:22 crc kubenswrapper[4706]: I1208 20:36:22.412205 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/pull/0.log" Dec 08 20:36:22 crc kubenswrapper[4706]: I1208 20:36:22.444533 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210smd24_9a62102a-4f69-4782-9153-f2b7a3136127/extract/0.log" Dec 08 20:36:22 crc kubenswrapper[4706]: I1208 20:36:22.643131 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/util/0.log" Dec 08 20:36:22 crc kubenswrapper[4706]: I1208 20:36:22.864385 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/pull/0.log" Dec 08 20:36:22 crc kubenswrapper[4706]: I1208 20:36:22.872823 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/util/0.log" Dec 08 20:36:22 crc kubenswrapper[4706]: I1208 20:36:22.935083 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/pull/0.log" Dec 08 20:36:23 crc kubenswrapper[4706]: I1208 20:36:23.183120 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/util/0.log" Dec 08 20:36:23 crc kubenswrapper[4706]: I1208 20:36:23.246691 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/extract/0.log" Dec 08 20:36:23 crc kubenswrapper[4706]: I1208 20:36:23.263754 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7b5aa1f5b38b68c96e281700110eb6f32773ca4b2682978fa6f2ffb2c1rm7nd_b0c2a5d5-da86-47ef-9785-323261350321/pull/0.log" Dec 08 20:36:23 crc kubenswrapper[4706]: I1208 20:36:23.428077 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/util/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.005823 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/pull/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.033374 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/util/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.035495 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/pull/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.215258 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/util/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.270152 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/pull/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.277357 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83hv8vf_07e73e83-48b7-43aa-b638-8c5898173956/extract/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.373794 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-utilities/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.572325 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-utilities/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.589487 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-content/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.589928 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-content/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.825338 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-utilities/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.836508 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-utilities/0.log" Dec 08 20:36:24 crc kubenswrapper[4706]: I1208 20:36:24.854201 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/extract-content/0.log" Dec 08 20:36:25 crc kubenswrapper[4706]: I1208 20:36:25.301401 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-utilities/0.log" Dec 08 20:36:25 crc kubenswrapper[4706]: I1208 20:36:25.320769 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-content/0.log" Dec 08 20:36:25 crc kubenswrapper[4706]: I1208 20:36:25.398916 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-content/0.log" Dec 08 20:36:25 crc kubenswrapper[4706]: I1208 20:36:25.562760 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mj8dj_4a2b5a45-e2a5-4793-b284-38116ab1b244/registry-server/0.log" Dec 08 20:36:25 crc kubenswrapper[4706]: I1208 20:36:25.623856 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-content/0.log" Dec 08 20:36:25 crc kubenswrapper[4706]: I1208 20:36:25.649794 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/extract-utilities/0.log" Dec 08 20:36:25 crc kubenswrapper[4706]: I1208 20:36:25.914596 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/3.log" Dec 08 20:36:25 crc kubenswrapper[4706]: I1208 20:36:25.967683 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-q4hlk_89875d36-6d95-4d46-9fc5-c18d8fccaae3/marketplace-operator/2.log" Dec 08 20:36:26 crc kubenswrapper[4706]: I1208 20:36:26.086984 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-utilities/0.log" Dec 08 20:36:26 crc kubenswrapper[4706]: I1208 20:36:26.441416 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-content/0.log" Dec 08 20:36:26 crc kubenswrapper[4706]: I1208 20:36:26.474647 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-content/0.log" Dec 08 20:36:26 crc kubenswrapper[4706]: I1208 20:36:26.533335 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-utilities/0.log" Dec 08 20:36:26 crc kubenswrapper[4706]: I1208 20:36:26.649251 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cqfh4_584d998e-f5d2-40eb-98d9-a3d5a6918144/registry-server/0.log" Dec 08 20:36:26 crc kubenswrapper[4706]: I1208 20:36:26.733233 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-utilities/0.log" Dec 08 20:36:26 crc kubenswrapper[4706]: I1208 20:36:26.832507 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/extract-content/0.log" Dec 08 20:36:26 crc kubenswrapper[4706]: I1208 20:36:26.907295 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s5pmf_c4f715e7-a6b1-4f7d-b1fe-cdc455423864/registry-server/0.log" Dec 08 20:36:27 crc kubenswrapper[4706]: I1208 20:36:27.024574 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-utilities/0.log" Dec 08 20:36:27 crc kubenswrapper[4706]: I1208 20:36:27.486611 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-content/0.log" Dec 08 20:36:27 crc kubenswrapper[4706]: I1208 20:36:27.528605 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-content/0.log" Dec 08 20:36:27 crc kubenswrapper[4706]: I1208 20:36:27.585680 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-utilities/0.log" Dec 08 20:36:27 crc kubenswrapper[4706]: I1208 20:36:27.759473 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-utilities/0.log" Dec 08 20:36:27 crc kubenswrapper[4706]: I1208 20:36:27.773657 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/extract-content/0.log" Dec 08 20:36:28 crc kubenswrapper[4706]: I1208 20:36:28.320456 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l2crj_da615802-59e8-4f61-9543-cd20ced42c39/registry-server/0.log" Dec 08 20:36:43 crc kubenswrapper[4706]: I1208 20:36:43.554225 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-fntdr_90a15b07-3ae8-4265-a92b-d9dd052211e2/prometheus-operator/0.log" Dec 08 20:36:43 crc kubenswrapper[4706]: I1208 20:36:43.704510 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-64d6644894-ctjxx_3f251b1b-6576-4421-aa76-04a26d0e1302/prometheus-operator-admission-webhook/0.log" Dec 08 20:36:43 crc kubenswrapper[4706]: I1208 20:36:43.796077 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-64d6644894-hdhxp_f543c2b7-0b00-458c-b4e5-a8c89cfaff5c/prometheus-operator-admission-webhook/0.log" Dec 08 20:36:43 crc kubenswrapper[4706]: I1208 20:36:43.959721 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-4xlqn_3cbb9b0b-a2c7-47f8-92d2-de4829687a45/operator/0.log" Dec 08 20:36:43 crc kubenswrapper[4706]: I1208 20:36:43.980100 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-fqdzc_a8e2aac6-1870-4ddc-8650-c31ded55337f/perses-operator/0.log" Dec 08 20:36:59 crc kubenswrapper[4706]: I1208 20:36:59.692165 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55474f77b6-v6n6q_b600c67d-b283-41ed-85f1-d19f3671f40f/kube-rbac-proxy/0.log" Dec 08 20:36:59 crc kubenswrapper[4706]: I1208 20:36:59.719419 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55474f77b6-v6n6q_b600c67d-b283-41ed-85f1-d19f3671f40f/manager/0.log" Dec 08 20:37:05 crc kubenswrapper[4706]: I1208 20:37:05.835840 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:37:05 crc kubenswrapper[4706]: I1208 20:37:05.836578 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:37:30 crc kubenswrapper[4706]: E1208 20:37:30.583452 4706 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:41700->38.102.83.113:43781: write tcp 38.102.83.113:41700->38.102.83.113:43781: write: broken pipe Dec 08 20:37:35 crc kubenswrapper[4706]: I1208 20:37:35.836555 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:37:35 crc kubenswrapper[4706]: I1208 20:37:35.837292 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:38:05 crc kubenswrapper[4706]: I1208 20:38:05.835788 4706 patch_prober.go:28] interesting pod/machine-config-daemon-kp47b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 20:38:05 crc kubenswrapper[4706]: I1208 20:38:05.836578 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 20:38:05 crc kubenswrapper[4706]: I1208 20:38:05.836643 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" Dec 08 20:38:05 crc kubenswrapper[4706]: I1208 20:38:05.837748 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040"} pod="openshift-machine-config-operator/machine-config-daemon-kp47b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 20:38:05 crc kubenswrapper[4706]: I1208 20:38:05.837816 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerName="machine-config-daemon" containerID="cri-o://554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" gracePeriod=600 Dec 08 20:38:06 crc kubenswrapper[4706]: I1208 20:38:06.524673 4706 generic.go:334] "Generic (PLEG): container finished" podID="8be6c92a-aa1d-4749-963b-49e360f63b97" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" exitCode=0 Dec 08 20:38:06 crc kubenswrapper[4706]: I1208 20:38:06.524729 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" event={"ID":"8be6c92a-aa1d-4749-963b-49e360f63b97","Type":"ContainerDied","Data":"554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040"} Dec 08 20:38:06 crc kubenswrapper[4706]: I1208 20:38:06.524779 4706 scope.go:117] "RemoveContainer" containerID="4d17c26bf612a2b2ecf64b164dc8622c2374233ba25c61a4ad2d539b54439cf0" Dec 08 20:38:06 crc kubenswrapper[4706]: E1208 20:38:06.526864 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:38:07 crc kubenswrapper[4706]: I1208 20:38:07.541721 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:38:07 crc kubenswrapper[4706]: E1208 20:38:07.542294 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:38:20 crc kubenswrapper[4706]: I1208 20:38:20.609404 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:38:20 crc kubenswrapper[4706]: E1208 20:38:20.610578 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:38:33 crc kubenswrapper[4706]: I1208 20:38:33.608026 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:38:33 crc kubenswrapper[4706]: E1208 20:38:33.608893 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:38:47 crc kubenswrapper[4706]: I1208 20:38:47.622437 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:38:47 crc kubenswrapper[4706]: E1208 20:38:47.623460 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:38:58 crc kubenswrapper[4706]: I1208 20:38:58.608860 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:38:58 crc kubenswrapper[4706]: E1208 20:38:58.610374 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:39:07 crc kubenswrapper[4706]: I1208 20:39:07.226861 4706 generic.go:334] "Generic (PLEG): container finished" podID="74631943-0724-4b1c-a8ba-8042bdd46713" containerID="ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23" exitCode=0 Dec 08 20:39:07 crc kubenswrapper[4706]: I1208 20:39:07.226985 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4gl9/must-gather-clghq" event={"ID":"74631943-0724-4b1c-a8ba-8042bdd46713","Type":"ContainerDied","Data":"ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23"} Dec 08 20:39:07 crc kubenswrapper[4706]: I1208 20:39:07.228559 4706 scope.go:117] "RemoveContainer" containerID="ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23" Dec 08 20:39:07 crc kubenswrapper[4706]: I1208 20:39:07.917400 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f4gl9_must-gather-clghq_74631943-0724-4b1c-a8ba-8042bdd46713/gather/0.log" Dec 08 20:39:10 crc kubenswrapper[4706]: I1208 20:39:10.608623 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:39:10 crc kubenswrapper[4706]: E1208 20:39:10.609393 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:39:20 crc kubenswrapper[4706]: I1208 20:39:20.346171 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f4gl9/must-gather-clghq"] Dec 08 20:39:20 crc kubenswrapper[4706]: I1208 20:39:20.347406 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-f4gl9/must-gather-clghq" podUID="74631943-0724-4b1c-a8ba-8042bdd46713" containerName="copy" containerID="cri-o://c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2" gracePeriod=2 Dec 08 20:39:20 crc kubenswrapper[4706]: I1208 20:39:20.364638 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f4gl9/must-gather-clghq"] Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.003240 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f4gl9_must-gather-clghq_74631943-0724-4b1c-a8ba-8042bdd46713/copy/0.log" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.004491 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.087468 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/74631943-0724-4b1c-a8ba-8042bdd46713-must-gather-output\") pod \"74631943-0724-4b1c-a8ba-8042bdd46713\" (UID: \"74631943-0724-4b1c-a8ba-8042bdd46713\") " Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.087673 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpp6h\" (UniqueName: \"kubernetes.io/projected/74631943-0724-4b1c-a8ba-8042bdd46713-kube-api-access-rpp6h\") pod \"74631943-0724-4b1c-a8ba-8042bdd46713\" (UID: \"74631943-0724-4b1c-a8ba-8042bdd46713\") " Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.333084 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74631943-0724-4b1c-a8ba-8042bdd46713-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "74631943-0724-4b1c-a8ba-8042bdd46713" (UID: "74631943-0724-4b1c-a8ba-8042bdd46713"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.402870 4706 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/74631943-0724-4b1c-a8ba-8042bdd46713-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.441532 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f4gl9_must-gather-clghq_74631943-0724-4b1c-a8ba-8042bdd46713/copy/0.log" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.441991 4706 generic.go:334] "Generic (PLEG): container finished" podID="74631943-0724-4b1c-a8ba-8042bdd46713" containerID="c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2" exitCode=143 Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.442069 4706 scope.go:117] "RemoveContainer" containerID="c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.442071 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4gl9/must-gather-clghq" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.476107 4706 scope.go:117] "RemoveContainer" containerID="ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.713001 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74631943-0724-4b1c-a8ba-8042bdd46713-kube-api-access-rpp6h" (OuterVolumeSpecName: "kube-api-access-rpp6h") pod "74631943-0724-4b1c-a8ba-8042bdd46713" (UID: "74631943-0724-4b1c-a8ba-8042bdd46713"). InnerVolumeSpecName "kube-api-access-rpp6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.813175 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpp6h\" (UniqueName: \"kubernetes.io/projected/74631943-0724-4b1c-a8ba-8042bdd46713-kube-api-access-rpp6h\") on node \"crc\" DevicePath \"\"" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.845228 4706 scope.go:117] "RemoveContainer" containerID="c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2" Dec 08 20:39:21 crc kubenswrapper[4706]: E1208 20:39:21.845912 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2\": container with ID starting with c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2 not found: ID does not exist" containerID="c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.845962 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2"} err="failed to get container status \"c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2\": rpc error: code = NotFound desc = could not find container \"c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2\": container with ID starting with c0b498ec51c63d70ff8d0a71cadee35fa0f471d4a407002a195d93333a7bbfd2 not found: ID does not exist" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.845990 4706 scope.go:117] "RemoveContainer" containerID="ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23" Dec 08 20:39:21 crc kubenswrapper[4706]: E1208 20:39:21.846467 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23\": container with ID starting with ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23 not found: ID does not exist" containerID="ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23" Dec 08 20:39:21 crc kubenswrapper[4706]: I1208 20:39:21.846495 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23"} err="failed to get container status \"ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23\": rpc error: code = NotFound desc = could not find container \"ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23\": container with ID starting with ac6c42821eb5054340a1ccbced40b472cc6c8f5e5baa39cad595ca8e1ae44f23 not found: ID does not exist" Dec 08 20:39:23 crc kubenswrapper[4706]: I1208 20:39:23.609527 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:39:23 crc kubenswrapper[4706]: E1208 20:39:23.613086 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:39:23 crc kubenswrapper[4706]: I1208 20:39:23.647386 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74631943-0724-4b1c-a8ba-8042bdd46713" path="/var/lib/kubelet/pods/74631943-0724-4b1c-a8ba-8042bdd46713/volumes" Dec 08 20:39:34 crc kubenswrapper[4706]: I1208 20:39:34.609199 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:39:34 crc kubenswrapper[4706]: E1208 20:39:34.610511 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:39:46 crc kubenswrapper[4706]: I1208 20:39:46.609288 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:39:46 crc kubenswrapper[4706]: E1208 20:39:46.610282 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:40:01 crc kubenswrapper[4706]: I1208 20:40:01.609295 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:40:01 crc kubenswrapper[4706]: E1208 20:40:01.639910 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:40:16 crc kubenswrapper[4706]: I1208 20:40:16.608840 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:40:16 crc kubenswrapper[4706]: E1208 20:40:16.611733 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:40:28 crc kubenswrapper[4706]: I1208 20:40:28.609527 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:40:28 crc kubenswrapper[4706]: E1208 20:40:28.612058 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:40:43 crc kubenswrapper[4706]: I1208 20:40:43.609889 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:40:43 crc kubenswrapper[4706]: E1208 20:40:43.611215 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:40:58 crc kubenswrapper[4706]: I1208 20:40:58.609351 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:40:58 crc kubenswrapper[4706]: E1208 20:40:58.610647 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:41:13 crc kubenswrapper[4706]: I1208 20:41:13.610023 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:41:13 crc kubenswrapper[4706]: E1208 20:41:13.611220 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:41:24 crc kubenswrapper[4706]: I1208 20:41:24.609943 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:41:24 crc kubenswrapper[4706]: E1208 20:41:24.611231 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.087235 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-52p6z"] Dec 08 20:41:35 crc kubenswrapper[4706]: E1208 20:41:35.089986 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="extract-content" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090016 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="extract-content" Dec 08 20:41:35 crc kubenswrapper[4706]: E1208 20:41:35.090039 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="extract-utilities" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090049 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="extract-utilities" Dec 08 20:41:35 crc kubenswrapper[4706]: E1208 20:41:35.090091 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74631943-0724-4b1c-a8ba-8042bdd46713" containerName="copy" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090099 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="74631943-0724-4b1c-a8ba-8042bdd46713" containerName="copy" Dec 08 20:41:35 crc kubenswrapper[4706]: E1208 20:41:35.090121 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerName="extract-content" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090129 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerName="extract-content" Dec 08 20:41:35 crc kubenswrapper[4706]: E1208 20:41:35.090147 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerName="registry-server" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090156 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerName="registry-server" Dec 08 20:41:35 crc kubenswrapper[4706]: E1208 20:41:35.090176 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerName="extract-utilities" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090183 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerName="extract-utilities" Dec 08 20:41:35 crc kubenswrapper[4706]: E1208 20:41:35.090198 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="registry-server" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090205 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="registry-server" Dec 08 20:41:35 crc kubenswrapper[4706]: E1208 20:41:35.090219 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74631943-0724-4b1c-a8ba-8042bdd46713" containerName="gather" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090225 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="74631943-0724-4b1c-a8ba-8042bdd46713" containerName="gather" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090560 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="495b7e1d-73b6-43d8-a009-314c3985d44e" containerName="registry-server" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090591 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="74631943-0724-4b1c-a8ba-8042bdd46713" containerName="gather" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090617 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="74631943-0724-4b1c-a8ba-8042bdd46713" containerName="copy" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.090637 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b37d3642-1357-4be7-8d17-1d0db472f060" containerName="registry-server" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.105834 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.110465 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-52p6z"] Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.219233 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-catalog-content\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.219629 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2ld6\" (UniqueName: \"kubernetes.io/projected/6615f930-5234-4c27-b999-5d74ce3ea026-kube-api-access-z2ld6\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.219708 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-utilities\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.324256 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-catalog-content\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.324374 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2ld6\" (UniqueName: \"kubernetes.io/projected/6615f930-5234-4c27-b999-5d74ce3ea026-kube-api-access-z2ld6\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.324406 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-utilities\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.324860 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-catalog-content\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.328762 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-utilities\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.391387 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2ld6\" (UniqueName: \"kubernetes.io/projected/6615f930-5234-4c27-b999-5d74ce3ea026-kube-api-access-z2ld6\") pod \"certified-operators-52p6z\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:35 crc kubenswrapper[4706]: I1208 20:41:35.442939 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:36 crc kubenswrapper[4706]: I1208 20:41:36.079136 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-52p6z"] Dec 08 20:41:36 crc kubenswrapper[4706]: I1208 20:41:36.160891 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52p6z" event={"ID":"6615f930-5234-4c27-b999-5d74ce3ea026","Type":"ContainerStarted","Data":"18eab730fc5277d208a3400d49d39c6420e9283a224883b493ede7fd50ae4615"} Dec 08 20:41:37 crc kubenswrapper[4706]: I1208 20:41:37.176178 4706 generic.go:334] "Generic (PLEG): container finished" podID="6615f930-5234-4c27-b999-5d74ce3ea026" containerID="6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264" exitCode=0 Dec 08 20:41:37 crc kubenswrapper[4706]: I1208 20:41:37.176414 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52p6z" event={"ID":"6615f930-5234-4c27-b999-5d74ce3ea026","Type":"ContainerDied","Data":"6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264"} Dec 08 20:41:37 crc kubenswrapper[4706]: I1208 20:41:37.179891 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 20:41:38 crc kubenswrapper[4706]: I1208 20:41:38.609205 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:41:38 crc kubenswrapper[4706]: E1208 20:41:38.610568 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:41:39 crc kubenswrapper[4706]: I1208 20:41:39.204764 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52p6z" event={"ID":"6615f930-5234-4c27-b999-5d74ce3ea026","Type":"ContainerStarted","Data":"e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368"} Dec 08 20:41:40 crc kubenswrapper[4706]: I1208 20:41:40.219367 4706 generic.go:334] "Generic (PLEG): container finished" podID="6615f930-5234-4c27-b999-5d74ce3ea026" containerID="e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368" exitCode=0 Dec 08 20:41:40 crc kubenswrapper[4706]: I1208 20:41:40.219460 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52p6z" event={"ID":"6615f930-5234-4c27-b999-5d74ce3ea026","Type":"ContainerDied","Data":"e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368"} Dec 08 20:41:41 crc kubenswrapper[4706]: I1208 20:41:41.239450 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52p6z" event={"ID":"6615f930-5234-4c27-b999-5d74ce3ea026","Type":"ContainerStarted","Data":"d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242"} Dec 08 20:41:41 crc kubenswrapper[4706]: I1208 20:41:41.281587 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-52p6z" podStartSLOduration=2.808353458 podStartE2EDuration="6.281566303s" podCreationTimestamp="2025-12-08 20:41:35 +0000 UTC" firstStartedPulling="2025-12-08 20:41:37.179651961 +0000 UTC m=+4779.821852954" lastFinishedPulling="2025-12-08 20:41:40.652864796 +0000 UTC m=+4783.295065799" observedRunningTime="2025-12-08 20:41:41.271959298 +0000 UTC m=+4783.914160311" watchObservedRunningTime="2025-12-08 20:41:41.281566303 +0000 UTC m=+4783.923767306" Dec 08 20:41:45 crc kubenswrapper[4706]: I1208 20:41:45.445426 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:45 crc kubenswrapper[4706]: I1208 20:41:45.446314 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:45 crc kubenswrapper[4706]: I1208 20:41:45.499769 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:46 crc kubenswrapper[4706]: I1208 20:41:46.352708 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:46 crc kubenswrapper[4706]: I1208 20:41:46.425140 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-52p6z"] Dec 08 20:41:48 crc kubenswrapper[4706]: I1208 20:41:48.314647 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-52p6z" podUID="6615f930-5234-4c27-b999-5d74ce3ea026" containerName="registry-server" containerID="cri-o://d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242" gracePeriod=2 Dec 08 20:41:48 crc kubenswrapper[4706]: I1208 20:41:48.968928 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.097594 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2ld6\" (UniqueName: \"kubernetes.io/projected/6615f930-5234-4c27-b999-5d74ce3ea026-kube-api-access-z2ld6\") pod \"6615f930-5234-4c27-b999-5d74ce3ea026\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.097809 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-utilities\") pod \"6615f930-5234-4c27-b999-5d74ce3ea026\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.097889 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-catalog-content\") pod \"6615f930-5234-4c27-b999-5d74ce3ea026\" (UID: \"6615f930-5234-4c27-b999-5d74ce3ea026\") " Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.098935 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-utilities" (OuterVolumeSpecName: "utilities") pod "6615f930-5234-4c27-b999-5d74ce3ea026" (UID: "6615f930-5234-4c27-b999-5d74ce3ea026"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.171338 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6615f930-5234-4c27-b999-5d74ce3ea026" (UID: "6615f930-5234-4c27-b999-5d74ce3ea026"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.201171 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.201223 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6615f930-5234-4c27-b999-5d74ce3ea026-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.202444 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6615f930-5234-4c27-b999-5d74ce3ea026-kube-api-access-z2ld6" (OuterVolumeSpecName: "kube-api-access-z2ld6") pod "6615f930-5234-4c27-b999-5d74ce3ea026" (UID: "6615f930-5234-4c27-b999-5d74ce3ea026"). InnerVolumeSpecName "kube-api-access-z2ld6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.304137 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2ld6\" (UniqueName: \"kubernetes.io/projected/6615f930-5234-4c27-b999-5d74ce3ea026-kube-api-access-z2ld6\") on node \"crc\" DevicePath \"\"" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.329938 4706 generic.go:334] "Generic (PLEG): container finished" podID="6615f930-5234-4c27-b999-5d74ce3ea026" containerID="d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242" exitCode=0 Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.330015 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52p6z" event={"ID":"6615f930-5234-4c27-b999-5d74ce3ea026","Type":"ContainerDied","Data":"d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242"} Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.330120 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52p6z" event={"ID":"6615f930-5234-4c27-b999-5d74ce3ea026","Type":"ContainerDied","Data":"18eab730fc5277d208a3400d49d39c6420e9283a224883b493ede7fd50ae4615"} Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.330149 4706 scope.go:117] "RemoveContainer" containerID="d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.330063 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52p6z" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.369641 4706 scope.go:117] "RemoveContainer" containerID="e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.386057 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-52p6z"] Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.399341 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-52p6z"] Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.405108 4706 scope.go:117] "RemoveContainer" containerID="6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.466511 4706 scope.go:117] "RemoveContainer" containerID="d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242" Dec 08 20:41:49 crc kubenswrapper[4706]: E1208 20:41:49.467174 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242\": container with ID starting with d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242 not found: ID does not exist" containerID="d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.467233 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242"} err="failed to get container status \"d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242\": rpc error: code = NotFound desc = could not find container \"d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242\": container with ID starting with d8a257a1abfdce9deac1eb712deb57fbcecf25885828c2f9a9c3066353e25242 not found: ID does not exist" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.467285 4706 scope.go:117] "RemoveContainer" containerID="e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368" Dec 08 20:41:49 crc kubenswrapper[4706]: E1208 20:41:49.467854 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368\": container with ID starting with e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368 not found: ID does not exist" containerID="e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.467935 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368"} err="failed to get container status \"e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368\": rpc error: code = NotFound desc = could not find container \"e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368\": container with ID starting with e6f77dc6cd1b3f09f6544aa42f19c05915c4813af851e938854e9018528bc368 not found: ID does not exist" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.467982 4706 scope.go:117] "RemoveContainer" containerID="6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264" Dec 08 20:41:49 crc kubenswrapper[4706]: E1208 20:41:49.468429 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264\": container with ID starting with 6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264 not found: ID does not exist" containerID="6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.468471 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264"} err="failed to get container status \"6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264\": rpc error: code = NotFound desc = could not find container \"6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264\": container with ID starting with 6a605998255bce20e4d10ef6b48c25a909714f82eb76b9eab7553d36bba0f264 not found: ID does not exist" Dec 08 20:41:49 crc kubenswrapper[4706]: I1208 20:41:49.622543 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6615f930-5234-4c27-b999-5d74ce3ea026" path="/var/lib/kubelet/pods/6615f930-5234-4c27-b999-5d74ce3ea026/volumes" Dec 08 20:41:51 crc kubenswrapper[4706]: I1208 20:41:51.609323 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:41:51 crc kubenswrapper[4706]: E1208 20:41:51.610167 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:42:03 crc kubenswrapper[4706]: I1208 20:42:03.609075 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:42:03 crc kubenswrapper[4706]: E1208 20:42:03.610359 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:42:18 crc kubenswrapper[4706]: I1208 20:42:18.608698 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:42:18 crc kubenswrapper[4706]: E1208 20:42:18.609783 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:42:32 crc kubenswrapper[4706]: I1208 20:42:32.608694 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:42:32 crc kubenswrapper[4706]: E1208 20:42:32.609853 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:42:46 crc kubenswrapper[4706]: I1208 20:42:46.609104 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:42:46 crc kubenswrapper[4706]: E1208 20:42:46.610242 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" Dec 08 20:43:00 crc kubenswrapper[4706]: I1208 20:43:00.608669 4706 scope.go:117] "RemoveContainer" containerID="554b57376c85a860728af30515b418a24debc032688bccf5744bc7c7472f5040" Dec 08 20:43:00 crc kubenswrapper[4706]: E1208 20:43:00.609876 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kp47b_openshift-machine-config-operator(8be6c92a-aa1d-4749-963b-49e360f63b97)\"" pod="openshift-machine-config-operator/machine-config-daemon-kp47b" podUID="8be6c92a-aa1d-4749-963b-49e360f63b97" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515115634141024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015115634141017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015115622214016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015115622214015454 5ustar corecore